Monday, December 16, 2019

AWS ElasticSearch ?

What is AWS ElasticSearch?


Elasticsearch is a search engine based on the Lucene library. It provides a distributed, multitenant-capable full-text search engine with an HTTP web interface and schema-free JSON documents.
Amazon Elasticsearch Service is a fully managed service that makes it easy for you to deploy, secure, and run Elasticsearch cost-effectively at scale

Elasticsearch Commands:
create indexes
curl -XPUT --header 'Content-Type: application/json' https://search-easyaws-wdxsl2qnhulmfbttlougfbgdj4.us-east-1.es.amazonaws.com/samples/_doc/1 -d '{ "name" : "easy"}'

list all indexes
curl -X GET 'https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/_cat/indices?v'

Query using URL parameters
curl -X GET https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/samples/_search?q=school:Harvard

list all docs in index
curl -X GET 'https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/sample/_search'

Health Check
curl  'https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/_cat/health?v'

Get Indexes
curl  'https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/_cat/indices?v'


delete index
curl -X DELETE 'https://search-esayaws-t4owdr3ym7zfiszspj6tdp526e.us-east-1.es.amazonaws.com/samples'


Tuesday, November 19, 2019

how to skip stage in jenkins pipeline


Skipped Stages in Jenkins Scripted Pipeline



To show all stages at every build even if not executed is a

good practice and brings transparency into pipelines

with conditional steps or stages. Add the when condition

to a stage of your pipeline:





pipeline {
    agent any
  stages {
    stage ('1 choice') {
      when {
                expression { choice == '1'}
            }
            steps {
                echo "Hello, Choice 1!"
            }
    }
    stage ('2 choice') {
      when {
                expression { choice == '2'}
            }
            steps {
                echo "Hello, Choice 2!"
            }
    }
  stage ('3 choice') {
      when {
                expression { choice == '3'}
            }
            steps {
                echo "Hello, Choice 3!"
            }
        }
  }
}







Monday, October 21, 2019

Python Class with Attribute Example


#################################################
# Python Class with Attribute Example
#################################################

class class_with_Attribute:
#class attribute Declaration
        type="Normal"

t1 = class_with_Attribute()
t2 = class_with_Attribute()

#Getting Output of Class Attribute
print (t1.type,t2.type)
#Output would be
# Normal Normal

#updating class Attribute value
testing.type="Advance"

print (t1.type,t2.type)
#Output would be
# Advance Advance

Thursday, October 17, 2019

What is Terragrunt ?

What is Terragrunt ?

Terragrunt is a thin wrapper for Terraform that provides extra tools for working with multiple Terraform modules. 



Terragrunt is a thin wrapper for Terraform that provides extra tools for keeping your Terraform configurations DRY, working with multiple Terraform modules, and managing remote state.

Get More details about terragrunt : https://github.com/gruntwork-io/terragrunt

Monday, October 14, 2019

Terraform - should I use user_data or remote-exec?

Terraform - should I use user_data or provisioner to bootstrap a resource?


You should use user_data. The user data field is idiomatic because it's native to AWS, whereas the remote-exec provisioner is specific to Terraform, which is just one of many ways to call the AWS API.
Also, the user-data is viewable in the AWS console, and often an important part of using Auto Scaling Groups in AWS, where you want each EC2 Instance to execute the same config code when it launches. It's not possible to do that with Terraform's remote-exec provisioner

Thursday, October 10, 2019

Route53 to S3 Redirection

Route53 to S3 Redirection
Route53 to S3 Redirection
Create Hosted Zone for the domain that you want to redirect to s3
Create S3 bucket with the same name as your domain

Note: The sites must use HTTP because the redirect 
can't connect to Amazon S3 over HTTPS.

Select the bucket you just created, and then choose
Properties. Expand the Static Website Hosting
drop-down menu, and then choose Redirect all
requests to another hostname.

In the Redirect all requests to another hostname
field, enter example.net, and then choose Save.

In the Route 53 console, select the hosted zone
named example.com.

Create a resource record for example.com with
the following values: Record Type: Choose A –
IPv4 address. Alias: Choose Yes. Alias Target:
Select the blank value field, expand the drop-down
menu, and then choose your S3 bucket under the
heading S3 Website Endpoints.

You should now be able to redirect your domain to s3.
you can confirm this by adding index.html file in s3 and
check with your DNS.

Amazon Route 53 Domain Configuration

Amazon Route 53

Domain Configuration

Configure DNS Records on Route 53

Create Hosted Zone for Public Hosting Domain as per below snaps

Domain to IP Redirection 

A Record Create

  1. Click Hosted Zones and select the domain you wish 
to configure.
  1. Click Create Record Set to add a new record, then select 
A - IPv4 address for the record type.
  1. Leave the Name field blank and enter the A record value
provided by Pantheon in the Value field.
  1. Select Desired Time to Live (TTL
  2. Click Create:

 

 

Domain to ELB Redirection 

Alias with A Records:

  1. Click Create Record Set to add a new record, then select 
A - IPv4 address for the record type.
  1. Leave the Name field blank and enter the A record value
provided by Pantheon in the Value field.
  1. Select Alias as Yes and select Load balance DNS in value
Field dropdown box
  1. Click Create:

AAAA Records

  1. Click Create Record Set to add a new record, then select
AAAA - IPv6 address for the record type.
  1. Leave the Name field blank and enter the both AAAA records
value provided by Pantheon in the Value field.
  1. Select desired TTL Value
  2. Click Create:
Route 53 AAAA records

A Record for subdomain

An A record is required to configure a subdomain (e.g., www.example.in).
  1. Click Create Record Set to add a new record, then select 
A - IPv4 address for the record type.
  1. Enter www in the Name field and enter the A record value 
provided by Pantheon in the Value field (e.g. ip address)
  1. Select desired Time to Live (TTL).
  2. Click Create:

AWS IAM EC2 Instance Role using Terraform

AWS IAM EC2 Instance Role using Terraform

IAM Roles are used to granting the application access
to AWS Services without using permanent credentials.
IAM Role is one of the safer ways to give permission to
your EC2 instances.
We can attach roles to an EC2 instance, and that allows us to give
permission to EC2 instance to use other AWS Services, for example,
S3 buckets, database dynamo DB, ASG
Create an Assume role as below, assume_role_policy
(Required) The policy that grants an entity permission
to assume the role.

resource "aws_iam_role" "ec2_role" {
name = "ec2_terraform_role"

assume_role_policy = <<EOF
{
"Version": "2012-10-17",
"Statement": [
  {
    "Action": "sts:AssumeRole",
    "Principal": {
      "Service": "ec2.amazonaws.com"
    },
    "Effect": "Allow",
    "Sid": ""
  }
]
}
EOF

tags = {
    tag-key = "tag-value"
}
}
This is going to create IAM role but we can’t link this role to
EC2 Instance and for that, we need EC2 instance Profile

Create EC2 Instance Profile using below code

resource "aws_iam_instance_profile" "ec2_profile" {
name = "ec2_terraform_profile"
role = "${aws_iam_role.ec2_role.name}"
}

Now if we execute the above code, we have Role and Instance
Profile but with no permission.so lets add IAM Policies which
allows EC2 instance to execute specific commands for example:
access to S3 Bucket
Adding IAM Policies,for giving full access to S3 bucket
resource "aws_iam_role_policy" "test_policy" {
name = "ec2_terraform_policy"
role = "${aws_iam_role.ec2_role.id}"

policy = <<EOF
{
"Version": "2012-10-17",
"Statement": [
  {
    "Action": [
      "s3:*"
    ],
    "Effect": "Allow",
    "Resource": "*"
  }
]
}
EOF
}
Attach this role to EC2 instance
resource "aws_instance" "ebs_instance_example" {
  ami           = lookup(var.ami_id, var.region)
  instance_type = var.instance_type
  iam_instance_profile = aws_iam_instance_profile.ec2_profile.name
  # key name
  key_name = var.key_name
  # User data passing through template rendering

  tags = {
    Name = "Roles with Ec2"
  }
}
It’s time to execute code run


terraform init
Let you see what terraform will do before making the actual changes
terraform plan
To actually create the instance we need to run terraform apply
terraform apply

Wireless Security Configuration: Protect Your Network Now!

Introduction: In today’s connected world, wireless networks are as common as smartphones, and they’re often the gateway to our personal, pr...