_This issue was originally opened by @boompig as hashicorp/terraform#8330. It was migrated here as part of the provider split. The original body of the issue is below._
Terraform v0.7.0
Probably all of AWS, observed with S3.
variable "region" {
default = "us-west-2"
}
provider "aws" {
region = "${var.region}"
profile = "fake_profile"
}
resource "aws_s3_bucket" "bucket" {
bucket = "fakebucket-something-test-1"
acl = "private"
}
https://gist.github.com/boompig/f05871140b928ae02b8f835d745158ac
Should successfully login then give "noop" text.
Does not read the correct profile from environment variable. Works if you provide the profile name in the file, though.
your_real_profile
terraform apply
I don't understand the "AWS_PROFILE" requirement. Why is this required to work, when it is not required in Packer, for example? Simply defining this in the aws{} block should be sufficient.
For example, the following does not work (0.9.6):
aws.tf
provider "aws" {
region = "us-east-1"
profile = "sandbox"
}
~/.aws/credentials
[sandbox]
aws_access_key_id = FOO
aws_secret_access_key = BAR
region = us-east-1
~/.aws/config
[default]
region = us-west-2
[profile sandbox]
# Nothing required here, see credentials file
Then running:
terraform plan
What does work? Same config but:
AWS_PROFILE=sandbox terraform plan
So, why does the first fail, while the second work? What is the point of the ENV variable?
_(also still trying to figure out how region fits into this whole thing, since it seems to be equally arbitrary)_
Finally, why is this whole profiles thing so janky? Like, just make it simple:
terraform -aws-profile=foo plan
and call it a day already :-\
Terraform v0.9.8
Im not able to recreate this issue.
My config:
variable "region" {
default = "us-east-1"
}
provider "aws" {
profile = "sub_account"
region = "${var.region}"
}
resource "aws_s3_bucket" "bucket" {
bucket = "fakebucket-something-test-6"
acl = "private"
}
TF_LOG=DEBUG terraform apply
It uses the key pair from sub_account profile
when I change the profile to test_account
It uses the test_account
credentials
You shouldn't need to set profile
in your provider and can just set theAWS_PROFILE
in your shell environment. At least that is working for me on 0.9.6
.
Sorry for the confusion on my end (and the noise). I realized I was missing something crucial. The initial state config and the subsequent terraform run are actually separate credentials. Profiles was working as expected the entire time, but I could not make the initial connection the state bucket with the default profile.
profile
is not working for me.
provider "aws" {
region = "us-east-2"
profile = "dev"
shared_credentials_file = "~/.aws/credentials"
}
$ terraform plan
Refreshing Terraform state in-memory prior to plan...
The refreshed state will be used to calculate this plan, but will not be
persisted to local or remote state storage.
Error refreshing state: 1 error(s) occurred:
* provider.aws: No valid credential sources found for AWS Provider.
Please see https://terraform.io/docs/providers/aws/index.html for more information on
providing credentials for the AWS Provider
Same result with:
export AWS_DEFAULT_PROFILE=dev
export AWS_PROFILE=dev
My environment:
$ terraform -version
Terraform v0.9.11
OS = macOS 10.12
@seanorama Do you use roles in your profile?
you do not need to set any environment variable and works out of the box.
Terraform 0.9.11
main.tf
variable "region" {
default = "us-east-1"
}
provider "aws" {
profile = "production"
region = "${var.region}"
}
resource "aws_s3_bucket" "bucket" {
bucket = "fakebucket-something-test-6"
acl = "private"
}
What works
~/.aws/config
[default]
region = eu-west-1
[profile test]
region = eu-west-1
[profile production]
region = eu-west-1
~/.aws/credentials
[default]
aws_access_key_id = foo
aws_secret_access_key = bar
[test]
aws_access_key_id = baz
aws_secret_access_key = blah
[production]
aws_access_key_id = boo
aws_secret_access_key = baa
What DOES NOT work
~/.aws/config
[default]
region = eu-west-1
[profile test]
region = eu-west-1
role_arn = some_role
source_profile = default
[profile production]
region = eu-west-1
role_arn = some_other_role
source_profile = default
~/.aws/credentials
[default]
aws_access_key_id = foo
aws_secret_access_key = bar
USING roles in profile?
This works!
variable "region" {
default = "us-east-1"
}
provider "aws" {
region = "${var.region}"
assume_role {
role_arn = "some_role"
}
}
resource "aws_s3_bucket" "bucket" {
bucket = "fakebucket-something-test-6"
acl = "private"
}
Just putting it here if someone else finds this problem.
profile works out of the box if you configured it correctly on awscli on awscli 1.11.113
and Terraform v0.10.4
aws configure --profile newprofile
provider "aws" {
region = "eu-west-2"
profile = "newprofile"
}
doesnt work for me. just tried on terraform version v0.11.3 and aws cli aws-cli/1.14.10
Ignore. look like i had an issue with my key. recreated and worked fine
Just in case, check out #1184 as well
The issue is that I'm already using AWS_PROFILE with Packer and boto3 and it works perfectly. To use Terraform I need to unset AWS_PROFILE AND add a profile in the Terraform provider config. This needs to be fixed ASAP - pick one of the other because this is overcomplicating the whole thing.
@kjenney I think they did. Check out #2883
Nope. I just verified against 1.15.0 and got the same exception. The assumed Role has AWSAdmin and boto3 will perform any operations with AWS_PROFILE set:
import boto3
import botocore
import os
os.environ['AWS_PROFILE'] = 'dev'
os.environ['AWS_DEFAULT_REGION'] = 'us-east-1'
ec2 = boto3.resource('ec2')
for i in ec2.instances.all(): print(i)
$ python listec2instances.py
ec2.Instance(id='i-0328bd472a76c42fb')
Here's my Terraform config:
provider "aws" {
region = "${var.aws_region}"
version = "~> 1.14"
}
$ export AWS_PROFILE="dev"
$ aws-cli/1.15.0 Python/3.6.5 Darwin/17.5.0 botocore/1.10.0
$ ls -altr .terraform/plugins/darwin_amd64/terraform-provider-aws_v1.15.0_x4
$ terraform plan
Refreshing Terraform state in-memory prior to plan...
The refreshed state will be used to calculate this plan, but will not be
persisted to local or remote state storage.
null_resource.default: Refreshing state... (ID: 1816969938248271300)
Error: Error refreshing state: 1 error(s) occurred:
* provider.aws: No valid credential sources found for AWS Provider.
Please see https://terraform.io/docs/providers/aws/index.html for more information on
providing credentials for the AWS Provider
So after digging further into #2883 I found that AWS_SDK_LOAD_CONFIG needs to be set for AWS_PROFILE to work. There is no mention of that in this issue OR in the public provider documentation: https://www.terraform.io/docs/providers/aws/index.html. This is an acceptable fix but it needs to be documented.
@kjenney yeah, that environment variable is poorly documented in all the SDKs. I have had to dig through each one to see if there was support for it. Sorry you didn't know about that, and sorry I didn't think to bring it up sooner :-/
Will there be a solution for the aws provider and an s3 backend that uses profiles with assumed roles? Seems like this might be causing some problems and unfortunately it is locked https://github.com/hashicorp/terraform/issues/13589.
Setting AWS_SDK_LOAD_CONFIG
and AWS_PROFILE
works with profiles that have aws_secret_access_key
and aws_access_key_id
, but does not work if the profile is setup like below:
[role]
role_arn = arn:aws:iam::{{AcountID}}:role/RoleName
source_profile = profile
Looking for any suggestions? My current solution is statically defining the profile in the provider and the backend, but it has already caused multiple problems :(
@alock The following configuration works for my team.
Our TF code has been converted to a single large module, with instances of this module for each environment. The directory layout is like this:
infra/project
/bastion.tf
/variables.tf
infra/staging
/main.tf
/provider_backend.tf
infra/prod
/main.tf
/provider_backend.tf
The main.tf
files above instantiate the module project
:
variable "region" {
default = "us-east-2"
}
module "staging" {
source = "../project"
region = "${var.region}"
}
The interesting part is provider_backend.tf
:
provider "aws" {
region = "${var.region}"
version = "~> 1.16"
assume_role {
role_arn = "arn:aws:iam::00000000000:role/OrganizationAccountAccessRole"
}
}
terraform {
backend "s3" {
region = "ap-southeast-2"
bucket = "project-terraform-states"
key = "project_staging"
}
}
This config can be used with the AWS_PROFILE
set to your identity account, and then terraform will assume the correct role before executing.
cd infra/staging
AWS_PROFILE=master terraform plan
as @kjenney said "AWS_SDK_LOAD_CONFIG needs to be set for AWS_PROFILE"
It seems like we can't reproduce this issue. To help the maintainers find the actionable issues in the tracker, I'm going to close this out, but if anyone is still experiencing this and can either supply a reproduction or logs, feel free to reply below or open a new issue. Thanks!
@paultyng - Maybe this wasn't meant to be solved with this issue, but I feel like terraform should be able to support environment variables without having to set explicit profiles
or assume_role
blocks in the files we commit. It can make it difficult since my team creates profiles-names uniquely and some have different Roles depending on the position. I thought that terraform would support env variables just like the AWSCLI. Below are some files that I used to test and highlight my problem. Everything wrapped in {{}}
are redacted variables.
~/.aws/config
[profile {{PROFILE_NAME}}]
output = json
region = us-west-2
role_arn = arn:aws:iam::{{ACCOUNT_NUM}}:role/{{ROLE_NAME}}
source_profile = {{SOURCE}}
[profile {{SOURCE}}]
output = json
region = us-west-2
~/.aws/credentials
[{{PROFILE_NAME}}]
role_arn = arn:aws:iam::{{ACCOUNT_NUM}}:role/{{ROLE_NAME}}
source_profile = {{SOURCE}}
[{{SOURCE}}]
aws_access_key_id = {{REMOVED}}
aws_secret_access_key = {{REMOVED}}
main.tf
provider "aws" {
region = "us-west-2"
}
terraform {
backend "s3" {
region = "us-west-2"
bucket = "{{S3_BUCKET}}"
key = "{{KEYNAME}}.tfstate"
encrypt = "true"
dynamodb_table = "{{TABLE_NAME}}"
acl = "bucket-owner-full-control"
}
}
Failing commands
$ AWS_PROFILE={{PROFILE_NAME}} aws iam list-account-aliases
{
"AccountAliases": [
"{{AWS_ALIAS}}"
]
}
$ AWS_PROFILE={{PROFILE_NAME}} AWS_SDK_LOAD_CONFIG=1 terraform init
Initializing the backend...
Error configuring the backend "s3": No valid credential sources found for AWS Provider.
Please see https://terraform.io/docs/providers/aws/index.html for more information on
providing credentials for the AWS Provider
Please update the configuration in your Terraform files to fix this error
then run this command again.
I've been able to reproduce this locally and have managed to diagnose the issue and worked out why the AWS backend wasn't able to pick up the AWS_PROFILE
where as the AWS provider was picking it up.
In short, the reason is that this issue is fixed in this repo but version of this repo that is present in the Terraform repo is an old version before this change was made.
@paultyng or someone at HashiCorp - can you update the version of the terraform-provider-aws library in the Terraform repo to fix this?
Looks like there's already a PR to cover this https://github.com/hashicorp/terraform/pull/17901
Upstream PR has been merged and will release with Terraform core 0.11.8.
So after digging further into #2883 I found that AWS_SDK_LOAD_CONFIG needs to be set for AWS_PROFILE to work. There is no mention of that in this issue OR in the public provider documentation: https://www.terraform.io/docs/providers/aws/index.html. This is an acceptable fix but it needs to be documented.
For anyone interested, this still hasn't been documented (and I just lost a lot of time because of the elusive AWS_SDK_LOAD_CONFIG), so I've opened a couple of PRs documenting it:
https://github.com/hashicorp/terraform/pull/21122
https://github.com/terraform-providers/terraform-provider-aws/pull/8451
Why is this closed? I have the same issue and nothing works. The configuration is fine with all aws cli, boto3 etc. so there is nothing wrong with .aws credentials. This sholud not be broken Setting the mysterious extra variable somehow to do do with a mumble GO something or other does nothing for me either. This is embarrassing guys.
I'm going to lock this issue because it has been closed for _30 days_ โณ. This helps our maintainers find and focus on the active issues.
If you feel this issue should be reopened, we encourage creating a new issue linking back to this one for added context. Thanks!
Most helpful comment
Finally, why is this whole profiles thing so janky? Like, just make it simple:
terraform -aws-profile=foo plan
and call it a day already :-\