Search

What is AWS CLI and How to Install it?

Whether you are a small business or a big organisation, you must be familiar with Amazon Web Services. AWS is the world’s most widely used cloud computing platform – something that lets you move faster, operate more securely and save substantial costs. AWS offers 165 services spanning a wide range including computing, tools for the Internet of Things, developer tools, deployment, analytics, database, networking, etc. The AWS Command Line Interface (CLI) is a unified tool to manage your AWS services.Do you want to Get AWS Certified? Learn about various  AWS Certification in detailWhat is AWS CLI?The AWS Command Line Interface (CLI) is a unified tool that manages the AWS services for you. You only have to download and configure one simple tool to control a plethora of AWS services. They are automated through scripts and help you implement a certain level of automation. It is indeed AWS CLI that makes AWS so dynamic and easy to use.How to Install AWS CLI?Putting AWS CLI to use in your AWS involves a few steps.There are a few different ways to install it – you need to choose what works for your system.Every operating system has a different method of installation.Some generic steps are followed after the installation.Then comes the configuration.You will also need to upgrade it timely. Follow these steps to successfully install and configure the AWS CLI for use.Ways to Install:You can effectively install the AWS Command Line Interface (AWS CLI) using:pipa virtual environmenta bundled installerWhat do you need?Unix, macOS, Linux, WindowsPython 3 version 3.3+ or Python 2 version 2.6.5+It is important to know that you may not be able to use an older version of Python with all AWS Services. Update to a newer version if there are Insecure Platform Warning or deprecation notices. To find out what version you currently have, visit:  https://github.com/aws/aws-cli/blob/master/CHANGELOG.rst.1. Installing the AWS CLI Using pipPip is the main distribution method for the AWS CLI on macOS, Windows and Linux. It is a package manager for Python.Installing the current AWS CLI VersionIf you have pip and a supported version of Python, use the following command to install the AWS CLI. Use the pip3 command if you have Python version 3+ installed:$ pip3 install awscli --upgrade –userThe --upgrade option commands pip3 to upgrade the requirements that are already installed. The --user option commands pip3 to install the program to a subdirectory of the user directory. Doing this avoids the complication of modifying libraries used by your operating system.Upgrading to the latest versionUse the pip list -o command to identify packages that are "outdated”:$ aws --version aws-cli/1.16.170 Python/3.7.3 Linux/4.14.123-111.109.amzn2.x86_64 botocore/1.12.160 $ pip3 list -o Package Version Latest Type ---------- -------- -------- ----- awscli 1.16.170 1.16.198 wheel botocore 1.12.160 1.12.188 wheelNow, run pip install --upgrade to get the latest version:$ pip3 install --upgrade --user awscli Collecting aws cli Downloadinghttps://files.pythonhosted.org/packages/dc/70/b32e9534c32fe9331801449e1f7eacba6a1992c2e4af9c82ac9116661d3b/awscli-1.16.198-py2.py3-none-any.whl (1.7MB) |████████████████████████████████| 1.7MB 1.6MB/s Collecting botocore==1.12.188 (from awscli) Using cached https://files.pythonhosted.org/packages/10/cb/8dcfb3e035a419f228df7d3a0eea5d52b528bde7ca162f62f3096a930472/botocore-1.12.188-py2.py3-none-any.whl Requirement already satisfied, skipping upgrade: docutils>=0.10 in ./venv/lib/python3.7/site-packages (from awscli) (0.14) Requirement already satisfied, skipping upgrade: rsa<=3.5.0,>=3.1.2 in ./venv/lib/python3.7/site-packages (from awscli) (3.4.2) Requirement already satisfied, skipping upgrade: colorama<=0.3.9,>=0.2.5 in ./venv/lib/python3.7/site-packages (from awscli) (0.3.9) Requirement already satisfied, skipping upgrade: PyYAML<=5.1,>=3.10; python_version != "2.6" in ./venv/lib/python3.7/site-packages (from awscli) (3.13) Requirement already satisfied, skipping upgrade: s3transfer<0.3.0,>=0.2.0 in ./venv/lib/python3.7/site-packages (from awscli) (0.2.0) Requirement already satisfied, skipping upgrade: jmespath<1.0.0,>=0.7.1 in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (0.9.4) Requirement already satisfied, skipping upgrade: urllib3<1.26,>=1.20; python_version >= "3.4" in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (1.24.3) Requirement already satisfied, skipping upgrade: python-dateutil<3.0.0,>=2.1; python_version >= "2.7" in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (2.8.0) Requirement already satisfied, skipping upgrade: pyasn1>=0.1.3 in ./venv/lib/python3.7/site-packages (from rsa<=3.5.0,>=3.1.2->awscli) (0.4.5) Requirement already satisfied, skipping upgrade: six>=1.5 in ./venv/lib/python3.7/site-packages (from python-dateutil<3.0.0,>=2.1; python_version >= "2.7"->botocore==1.12.188->awscli) (1.12.0) Installing collected packages: botocore, awscli Found existing installation: botocore 1.12.160 Uninstalling botocore-1.12.160: Successfully uninstalled botocore-1.12.160 Found existing installation: awscli 1.16.170 Uninstalling awscli-1.16.170: Successfully uninstalled awscli-1.16.170 Successfully installed awscli-1.16.198 botocore-1.12.1882. Installing the AWS CLI in a Virtual EnvironmentAnother option is to install the AWS CLI in a virtual environment to separate the tool and its dependencies. You can also use a different Python version for this purpose.3. Installing the AWS CLI Using an InstallerUse the bundled installer for automated and offline installation on Unix, macOS, and Linux. It includes the AWS CLI, its dependencies, and a shell script that is responsible for the installation. For Windows, try MSI installer. Both these methods simplify the initial installation. Installation of Linux or UnixBoth the platforms have identical installation process. You need to have Python’s latest version. We recommend using the bundled installer for this. The steps are as follows:1. To begin the installation:curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip"2. Unzip the downloaded package:unzip awscli-bundle.zip3. Run the installation:sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/awsUsing the -b option allows you to use the AWS CLI from any directory.Installation on Amazon LinuxThe AWS Command Line Interface comes preinstalled on both Amazon Linux and Amazon Linux 2. Below are the steps to install:1. Identify currently installed version:$ aws --version aws-cli/1.16.116 Python/3.6.8 Linux/4.14.77-81.59.amzn2.x86_64 botocore/1.12.1062. Use pip3 to install the latest version of the AWS CLI. If you run the command from within a Python virtual environment (venv), then you don't need to use the --user option.$ pip3 install --upgrade --user awscli3. Add the install location to the beginning of the PATH variable.$ export PATH=/home/ec2-user/.local/bin:$PATH4. Verify that you're running new version with aws --version.$ aws --version aws-cli/1.16.116 Python/3.6.8 Linux/4.14.77-81.59.amzn2.x86_64 botocore/1.12.106Installation on WindowsThe AWS Command Line Interface can be installed on Windows by using a standalone installer or through a pip - a package manager for Python> Through InstallerDownload the appropriate MSI installer.Run the downloaded MSI installer or the setup file.Follow these instructions:By default, the CLI installs to C:\Program Files\Amazon\AWSCLI (64-bit version) or C:\Program Files (x86)\Amazon\AWSCLI (32-bit version). To confirm the installation, use the aws --version command at a command prompt (open the Start menu and search for cmd to start a command prompt).C:\> aws --version aws-cli/1.16.116 Python/3.6.8 Windows/10 botocore/1.12.106If Windows is unable to find the program, you might need to close and reopen the command prompt to refresh the path, or add the installation directory to your PATH environment variable manually.> Through Pip1. Open Start menu→Command Prompt2. Verify that both Python and pip are installed correctly:C:\> python --version Python 3.7.1 C:\> pip3 --version pip 18.1 from c:\program files\python37\lib\site-packages\pip (python 3.7)3. Install AWS CLI via pipC:\> pip3 install awscli4. Check if the installation went rightC:\> aws --version aws-cli/1.16.116 Python/3.6.8 Windows/10 botocore/1.12.106To upgrade to the latest version:C:\> pip3 install --user --upgrade awscliInstallation on MAC OS> Through Installer1. Download the  AWS CLI Bundled Installer$ curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip"2. Unzip the package$ unzip awscli-bundle.zip3. Run the installation$ sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/awsThis command installs the AWS CLI to /usr/local/aws and creates the symlink aws in the /usr/local/bin directory. Using the -b option to create a symlink eliminates the need to specify the install directory in the user's $PATH variable. It enables users to run the AWS CLI by typing “aws” from any directory.If you want to see an explanation of the -i and -b options, use the -h option$ ./awscli-bundle/install -hthe commands summarized for easy cut and paste at the command line.curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip" unzip awscli-bundle.zip sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/aws> Through PIP1. Download and install the latest version of Python from Python.org.2. Download and run the pip3 installation script provided by the Python Packaging Authority$ curl -O https://bootstrap.pypa.io/get-pip.py $ python3 get-pip.py --user3. Use pip3 to install the AWS CLI. We recommend using the pip3 command if you use Python version 3+$ pip3 install awscli --upgrade --user4. See if AWS CLI is installed correctly$ aws --version AWS CLI 1.16.116 (Python 3.6.8)To upgrade to the latest version, run the command:$ pip3 install awscli --upgrade --userInstallation on Ubuntu> Through APT Package Manager1. Update the package repository cache$ sudo apt-get update2. Install AWS CLI with the following command$ sudo apt-get install awsclipress y and then press <Enter> to continue. Your screen should look something like this: 3. Now that it’s installed, check if it’s working properly or not$ aws --version> Through PIPAWS CLI being a Python module itself makes it easy for users who install it through PIP to update it on a regular basis. Assuming you have Python 3, follow the below steps to install:1. Install Python PIP with the following command$ sudo apt-get install python3-pipPress y and then press <Enter> to continue2. Install AWS CLI using PIP with the following command$ pip3 install awscli --upgrade --user3. Run AWS CLI with the following command$ python3 -m awscli --versionAfter InstallationAfter you have successfully installed AWS CLI, you need to set the Path to Include the AWS CLI in your system.> LINUXFind out the folder in which pip installed the AWS CLI$ which aws /home/username/.local/bin/awsYou can reference this as “~/.local/bin/” because of the reason that “/home/username” corresponds to ~ in Linux OSIn case you don't know where Python is installed, run this command$ which python /usr/local/bin/pythonIf this is the same folder you added to the path while installing pip, there’s nothing else to be done. Otherwise, perform those same steps again, adding this additional folder to the path.> WINDOWSThe Windows System PATH tells your PC where it can find specific directories:C:\> where awsC:\Program Files\Amazon\AWSCLI\bin\aws.exeFind out where the aws program is installedC:\> where c:\ awsC:\Program Files\Python37\Scripts\awsIf the command returns the following error, then it is not in the system PATH and you can't run it by typing its name.C:\> where c:\ awsINFO:Could not find files for the given pattern.In that case, you need to add the path manually. First, you need to search where it is installed on your computer:C:\> where /R c:\ awsc:\Program Files\Amazon\AWSCLI\bin\aws.exec:\Program Files\Amazon\AWSCLI\bincompat\aws.cmdc:\Program Files\Amazon\AWSCLI\runtime\Scripts\awsc:\Program Files\Amazon\AWSCLI\runtime\Scripts\aws.cmd...To modify your PATH variable (Windows)Press the Windows key and enter environment variables.Choose the Edit environment variables for your account.Choose the PATH →Edit.Add the path to the Variable value field. For example: C:\new\pathClick OK twice to apply the new settings.Close any running command prompts and reopen the command prompt window.> MAC OSLocate Python$ which python /usr/local/bin/pythonThe output might be the path to a symlink, not the actual program. Run ls -al to see where it points.$ ls -al /usr/local/bin/python ~/Library/Python/3.7/bin/python3.6Pip install programs in the same folder as the Python application. Add this folder to your PATH variable.To modify the PATH variable for macOS (and Linus or Unix):1. Find the shell profile script in the user folder. In case you don’t know which shell you have, run echo $SHELL2. Through the following, add an export command to the profile scriptexport PATH=~/.local/bin:$PATHThis adds a path, ~/.local/bin in this example, to the current PATH variable.3. The updated profile can now be loaded into your current session$ source ~/.bash_profile

What is AWS CLI and How to Install it?

10K
  • by Joydip Kumar
  • 10th Sep, 2019
  • Last updated on 11th Mar, 2021
  • 14 mins read
What is AWS CLI and How to Install it?

Whether you are a small business or a big organisation, you must be familiar with Amazon Web Services. AWS is the world’s most widely used cloud computing platform – something that lets you move faster, operate more securely and save substantial costs. AWS offers 165 services spanning a wide range including computing, tools for the Internet of Things, developer tools, deployment, analytics, database, networking, etc. The AWS Command Line Interface (CLI) is a unified tool to manage your AWS services.

Do you want to Get AWS Certified? Learn about various  AWS Certification in detail

What is AWS CLI?

The AWS Command Line Interface (CLI) is a unified tool that manages the AWS services for you. You only have to download and configure one simple tool to control a plethora of AWS services. They are automated through scripts and help you implement a certain level of automation. It is indeed AWS CLI that makes AWS so dynamic and easy to use.

How to Install AWS CLI?

Putting AWS CLI to use in your AWS involves a few steps.

  • There are a few different ways to install it – you need to choose what works for your system.
  • Every operating system has a different method of installation.
  • Some generic steps are followed after the installation.
  • Then comes the configuration.
  • You will also need to upgrade it timely. 

Follow these steps to successfully install and configure the AWS CLI for use.

Ways to Install:

You can effectively install the AWS Command Line Interface (AWS CLI) using:

Ways to install the AWS Command Line Interface in AWS CLI

  1. pip
  2. a virtual environment
  3. a bundled installer

What do you need?

  • Unix, macOS, Linux, Windows
  • Python 3 version 3.3+ or Python 2 version 2.6.5+

It is important to know that you may not be able to use an older version of Python with all AWS Services. Update to a newer version if there are Insecure Platform Warning or deprecation notices. To find out what version you currently have, visit:

  https://github.com/aws/aws-cli/blob/master/CHANGELOG.rst.

1. Installing the AWS CLI Using pip

Pip is the main distribution method for the AWS CLI on macOS, Windows and Linux. It is a package manager for Python.

Installing the current AWS CLI Version

If you have pip and a supported version of Python, use the following command to install the AWS CLI. Use the pip3 command if you have Python version 3+ installed:

$ pip3 install awscli --upgrade –user

The --upgrade option commands pip3 to upgrade the requirements that are already installed. The --user option commands pip3 to install the program to a subdirectory of the user directory. Doing this avoids the complication of modifying libraries used by your operating system.

Upgrading to the latest version

Use the pip list -o command to identify packages that are "outdated”:

$ aws --version
aws-cli/1.16.170 Python/3.7.3 Linux/4.14.123-111.109.amzn2.x86_64 botocore/1.12.160

$ pip3 list -o

Package    Version      Latest        Type 
----------     --------      --------        -----
awscli       1.16.170    1.16.198   wheel
botocore   1.12.160    1.12.188   wheel

Now, run pip install --upgrade to get the latest version:

$ pip3 install --upgrade --user awscli
Collecting  aws cli 
Downloadinghttps://files.pythonhosted.org/packages/dc/70/b32e9534c32fe9331801449e1f7eacba6a1992c2e4af9c82ac9116661d3b/awscli-1.16.198-py2.py3-none-any.whl (1.7MB)
     |████████████████████████████████| 1.7MB 1.6MB/s 
Collecting botocore==1.12.188 (from awscli)
Using cached https://files.pythonhosted.org/packages/10/cb/8dcfb3e035a419f228df7d3a0eea5d52b528bde7ca162f62f3096a930472/botocore-1.12.188-py2.py3-none-any.whl
Requirement already satisfied, skipping upgrade: docutils>=0.10 in ./venv/lib/python3.7/site-packages (from awscli) (0.14)
Requirement already satisfied, skipping upgrade: rsa<=3.5.0,>=3.1.2 in ./venv/lib/python3.7/site-packages (from awscli) (3.4.2)
Requirement already satisfied, skipping upgrade: colorama<=0.3.9,>=0.2.5 in ./venv/lib/python3.7/site-packages (from awscli) (0.3.9)
Requirement already satisfied, skipping upgrade: PyYAML<=5.1,>=3.10; python_version != "2.6" in ./venv/lib/python3.7/site-packages (from awscli) (3.13)
Requirement already satisfied, skipping upgrade: s3transfer<0.3.0,>=0.2.0 in ./venv/lib/python3.7/site-packages (from awscli) (0.2.0)
Requirement already satisfied, skipping upgrade: jmespath<1.0.0,>=0.7.1 in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (0.9.4)
Requirement already satisfied, skipping upgrade: urllib3<1.26,>=1.20; python_version >= "3.4" in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (1.24.3)
Requirement already satisfied, skipping upgrade: python-dateutil<3.0.0,>=2.1; python_version >= "2.7" in ./venv/lib/python3.7/site-packages (from botocore==1.12.188->awscli) (2.8.0)
Requirement already satisfied, skipping upgrade: pyasn1>=0.1.3 in ./venv/lib/python3.7/site-packages (from rsa<=3.5.0,>=3.1.2->awscli) (0.4.5)
Requirement already satisfied, skipping upgrade: six>=1.5 in ./venv/lib/python3.7/site-packages (from python-dateutil<3.0.0,>=2.1; python_version >= "2.7"->botocore==1.12.188->awscli) (1.12.0)
Installing collected packages: botocore, awscli
Found existing installation: botocore 1.12.160
Uninstalling botocore-1.12.160:
Successfully uninstalled botocore-1.12.160
Found existing installation: awscli 1.16.170
Uninstalling awscli-1.16.170:
Successfully uninstalled awscli-1.16.170
Successfully installed awscli-1.16.198 botocore-1.12.188

2. Installing the AWS CLI in a Virtual Environment

Another option is to install the AWS CLI in a virtual environment to separate the tool and its dependencies. You can also use a different Python version for this purpose.

3. Installing the AWS CLI Using an Installer

Use the bundled installer for automated and offline installation on Unix, macOS, and Linux. It includes the AWS CLI, its dependencies, and a shell script that is responsible for the installation. For Windows, try MSI installer. Both these methods simplify the initial installation. 

Installation of Linux or Unix

Both the platforms have identical installation process. You need to have Python’s latest version. We recommend using the bundled installer for this. The steps are as follows:

1. To begin the installation:

curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip"

2. Unzip the downloaded package:

unzip awscli-bundle.zip

3. Run the installation:

sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/aws

Using the -b option allows you to use the AWS CLI from any directory.

Installation on Amazon Linux

The AWS Command Line Interface comes preinstalled on both Amazon Linux and Amazon Linux 2. Below are the steps to install:

1. Identify currently installed version:

$ aws --version
aws-cli/1.16.116 Python/3.6.8 Linux/4.14.77-81.59.amzn2.x86_64 botocore/1.12.106

2. Use pip3 to install the latest version of the AWS CLI. If you run the command from within a Python virtual environment (venv), then you don't need to use the --user option.

$ pip3 install --upgrade --user awscli

3. Add the install location to the beginning of the PATH variable.

$ export PATH=/home/ec2-user/.local/bin:$PATH

4. Verify that you're running new version with aws --version.

$ aws --version
aws-cli/1.16.116 Python/3.6.8 Linux/4.14.77-81.59.amzn2.x86_64 botocore/1.12.106

Installation on Windows

The AWS Command Line Interface can be installed on Windows by using a standalone installer or through a pip - a package manager for Python

> Through Installer

  1. Download the appropriate MSI installer.
  2. Run the downloaded MSI installer or the setup file.
  3. Follow these instructions:

By default, the CLI installs to C:\Program Files\Amazon\AWSCLI (64-bit version) or C:\Program Files (x86)\Amazon\AWSCLI (32-bit version). To confirm the installation, use the aws --version command at a command prompt (open the Start menu and search for cmd to start a command prompt).

C:\> aws --version
aws-cli/1.16.116 Python/3.6.8 Windows/10 botocore/1.12.106

If Windows is unable to find the program, you might need to close and reopen the command prompt to refresh the path, or add the installation directory to your PATH environment variable manually.

> Through Pip

1. Open Start menu→Command Prompt
2. Verify that both Python and pip are installed correctly:

C:\> python --version
Python 3.7.1
C:\> pip3 --version
pip 18.1 from c:\program files\python37\lib\site-packages\pip (python 3.7)

3. Install AWS CLI via pip

C:\> pip3 install awscli

4. Check if the installation went right

C:\> aws --version
aws-cli/1.16.116 Python/3.6.8 Windows/10 botocore/1.12.106

To upgrade to the latest version:

C:\> pip3 install --user --upgrade awscli

Installation on MAC OS

> Through Installer

1. Download the  AWS CLI Bundled Installer

$ curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip"

2. Unzip the package

$ unzip awscli-bundle.zip

3. Run the installation

$ sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/aws

This command installs the AWS CLI to /usr/local/aws and creates the symlink aws in the /usr/local/bin directory. Using the -b option to create a symlink eliminates the need to specify the install directory in the user's $PATH variable. It enables users to run the AWS CLI by typing “aws” from any directory.

If you want to see an explanation of the -i and -b options, use the -h option

$ ./awscli-bundle/install -h

the commands summarized for easy cut and paste at the command line.

curl "https://s3.amazonaws.com/aws-cli/awscli-bundle.zip" -o "awscli-bundle.zip"
unzip awscli-bundle.zip
sudo ./awscli-bundle/install -i /usr/local/aws -b /usr/local/bin/aws

> Through PIP

1. Download and install the latest version of Python from Python.org.
2. Download and run the pip3 installation script provided by the Python Packaging Authority

$ curl -O https://bootstrap.pypa.io/get-pip.py
$ python3 get-pip.py --user

3. Use pip3 to install the AWS CLI. We recommend using the pip3 command if you use Python version 3+

$ pip3 install awscli --upgrade --user

4. See if AWS CLI is installed correctly

$ aws --version
AWS CLI 1.16.116 (Python 3.6.8)

To upgrade to the latest version, run the command:

$ pip3 install awscli --upgrade --user

Installation on Ubuntu

> Through APT Package Manager

1. Update the package repository cache

$ sudo apt-get update

2. Install AWS CLI with the following command

$ sudo apt-get install awscli

press y and then press <Enter> to continue. Your screen should look something like this: 



3. Now that it’s installed, check if it’s working properly or not

$ aws --version

> Through PIP

AWS CLI being a Python module itself makes it easy for users who install it through PIP to update it on a regular basis. Assuming you have Python 3, follow the below steps to install:

1. Install Python PIP with the following command

$ sudo apt-get install python3-pip

Press y and then press <Enter> to continue
2. Install AWS CLI using PIP with the following command

$ pip3 install awscli --upgrade --user

3. Run AWS CLI with the following command

$ python3 -m awscli --version

After Installation

After you have successfully installed AWS CLI, you need to set the Path to Include the AWS CLI in your system.

> LINUX

Find out the folder in which pip installed the AWS CLI

$ which aws
/home/username/.local/bin/aws

You can reference this as “~/.local/bin/” because of the reason that “/home/username” corresponds to ~ in Linux OS
In case you don't know where Python is installed, run this command

$ which python
/usr/local/bin/python

If this is the same folder you added to the path while installing pip, there’s nothing else to be done. Otherwise, perform those same steps again, adding this additional folder to the path.

> WINDOWS

The Windows System PATH tells your PC where it can find specific directories:

C:\> where aws

C:\Program Files\Amazon\AWSCLI\bin\aws.exe
Find out where the aws program is installed

C:\> where c:\ aws
C:\Program Files\Python37\Scripts\aws

If the command returns the following error, then it is not in the system PATH and you can't run it by typing its name.

C:\> where c:\ aws

INFO:Could not find files for the given pattern.

In that case, you need to add the path manually. First, you need to search where it is installed on your computer:

C:\> where /R c:\ aws
c:\Program Files\Amazon\AWSCLI\bin\aws.exe
c:\Program Files\Amazon\AWSCLI\bincompat\aws.cmd
c:\Program Files\Amazon\AWSCLI\runtime\Scripts\aws
c:\Program Files\Amazon\AWSCLI\runtime\Scripts\aws.cmd
...
To modify your PATH variable (Windows)

  1. Press the Windows key and enter environment variables.
  2. Choose the Edit environment variables for your account.
  3. Choose the PATH →Edit.
  4. Add the path to the Variable value field. For example: C:\new\path
  5. Click OK twice to apply the new settings.
  6. Close any running command prompts and reopen the command prompt window.

> MAC OS

Locate Python

$ which python
/usr/local/bin/python

The output might be the path to a symlink, not the actual program. Run ls -al to see where it points.

$ ls -al /usr/local/bin/python
~/Library/Python/3.7/bin/python3.6

Pip install programs in the same folder as the Python application. Add this folder to your PATH variable.

To modify the PATH variable for macOS (and Linus or Unix):

1. Find the shell profile script in the user folder. In case you don’t know which shell you have, run echo $SHELL

2. Through the following, add an export command to the profile script

export PATH=~/.local/bin:$PATH

This adds a path, ~/.local/bin in this example, to the current PATH variable.

3. The updated profile can now be loaded into your current session

$ source ~/.bash_profile

Joydip

Joydip Kumar

Solution Architect

Joydip is passionate about building cloud-based applications and has been providing solutions to various multinational clients. Being a java programmer and an AWS certified cloud architect, he loves to design, develop, and integrate solutions. Amidst his busy work schedule, Joydip loves to spend time on writing blogs and contributing to the opensource community.


Website : https://geeks18.com/

Join the Discussion

Your email address will not be published. Required fields are marked *

Suggested Blogs

A Glimpse Of The Major Leading SAFe® Versions

A Quick view of SAFe® Agile has gained popularity in recent years, and with good reason. Teams love this approach that allows them to get a value to the customer faster while learning and adjusting to change as needed. But teams often don’t work in isolation. Many teams work in the context of larger organizations.  Often Agile doesn’t fit their needs. Some teams need an Agile approach that scales to larger projects that involve multiple teams.   It’s possible to do this. That’s where the Scaled Agile Framework, or SAFe®, can help.Why SAFe® is the best scalable framework?The Scaled Agile Framework is a structured Agile approach for large enterprises. It’s prescriptive and provides a path for interdependent teams to gain the benefits of using an Agile approach.Scaled Agile provides guidance not only at the team level but also at the Program and Portfolio levels. It also has built-in coordinated planning across related teams who are working in Release Trains.These planning increments allow teams to plan together to work with customers and release value frequently in a way that’s sustainable to teams.And it supports continuous improvement.It’s a great way for large companies to maintain structure and roll out Agile at a large scale.  What is SAFe® 4.5? Scaled Agile, otherwise known as SAFe®, was initially released in 2011 by Dean Leffingwell as a knowledge base for enterprises to adopt Agile. Over the years it has grown and evolved. SAFe® 4.5 was released on June 22, 2017, to accommodate improvements to the framework. Following are some of the key improvements in SAFe® 4.5:Essential SAFe® and ConfigurabilityInnovation with Lean Startup and Lean UXScalable DevOps and Continuous DeliveryImplementation roadmapBenefits of SAFe® 4.5 to companies:Organizations who adopt SAFe® 4.5 will be able to gain the following benefits:1) Test ideas more quickly. SAFe® 4.5 has a build-in iterative development and testing. This lets teams get faster feedback to learn and adjust more quickly.2) Deliver much faster. The changes to SAFe® 4.5 allow teams to move complex work through the pipeline and deliver value to the customer faster.3) Simplify governance and improve portfolio performance. Guidance and support have been added at the Portfolio level to guide organizations in addressing Portfolio-level concerns in a scaled agile context. SAFe® 4.5 - Key areas of improvements:A. Essential SAFe® and ConfigurabilityFour configurations of SAFe® that provide a more configurable and scalable approach:Essential SAFe®: The most basic level that teams can use. It contains just the essentials that a team needs to get the benefits of SAFe®.Portfolio SAFe®: For enterprises that implement multiple solutions that have portfolio responsibilities such as governance, strategy, and portfolio funding.Large Solution: Complex solutions that involve multiple Agile Release Trains. These initiatives don’t require Portfolio concerns, but only include the Large Solution and Essential SAFe® elements.  SAFe® Full SAFe®: The most comprehensive level that can be applied to huge enterprise initiatives requiring hundreds of people to complete.Because SAFe® is a framework, that provides the flexibility to choose the level of SAFe® that best fits your organization’s needs.B. Innovation with Lean Startup and Lean UXRather than creating an entire project plan up-front, SAFe® teams focus on features. They create a hypothesis about what a new feature will deliver and then use an iterative approach to develop and test their hypothesis along the way. As teams move forward through development, they perform this development and test approach repeatedly and adjust as needed, based on feedback. Teams also work closely with end users to identify the Minimum Viable Product (MVP) to focus on first. They identify what will be most valuable to the customer most immediately. Then they rely on feedback and learning as they develop the solution incrementally. They adjust as needed to incorporate what they’ve learned into the features. This collaboration and fast feedback and adjustment cycle result in a more successful product.  C. Scalable DevOps & Continuous DeliveryThe addition of a greater focus on DevOps allows teams to innovate faster. Like Agile, DevOps is a mindset. And like Agile, it allows teams to learn, adjust, and deliver value to users incrementally. The continuous delivery pipeline allows teams to move value through the pipeline faster through continuous exploration, continuous integration, continuous deployment, and released on demand. DevOps breaks down silos and supports Agile teams to work together more seamlessly. This results in more efficient delivery of value to the end users faster. It’s a perfect complement to Scaled Agile.D. Implementation RoadmapSAFe® now offers a suggested roadmap to SAFe® adoption. While change can be challenging, the implementation roadmap provides guidance that can help with that organizational change.Critical Role of the SAFe® Program ConsultantSAFe® Program Consultants, or SPCs, are critical change agents in the transition to Scaled Agile.Because of the depth of knowledge required to gain SPC certification, they’re perfectly positioned to help the organization move through challenges of change.They can train and coach all levels of SAFe® participants, from team members to executive leaders. They can also train the Scrum Master, Product Owners, and Agile Release Train Engineers, which are critical roles in SAFe®.The SPC can also train teams and help them launch their Agile Release Trains (ARTs).And they can support teams on the path to continued improvement as they continue to learn and grow.The SPC can also help identify value streams in the organization that may be ready to launch Agile Release Trains.The can also help develop rollout plans for SAFe® in the enterprise.Along with this, they can provide important communications that help the enterprise understand the drivers and value behind the SAFe® transition.       How SAFe® 4.5 is backward compatible with SAFe® 4.0?Even if your organization has already adopted SAFe® 4.0, SAFe® 4.5 has been developed in a way that can be easily adopted without disruption. Your organization can adopt the changes at the pace that works best.Few Updates in the new courseware The courseware for SAFe® 4.5 has incorporated changes to support the changes in SAFe® 4.5.They include Implementing SAFe®, Leading SAFe®, and SAFe® for Teams.Some of the changes you’ll see are as follows:Two new lessons for Leading SAFe®Student workbookTrainer GuideNew look and feelUpdated LPM contentSmoother lesson flowNEW Course Delivery Enablement (CDE) Changes were made to improve alignment between SAFe® and Scrum:Iteration Review: Increments previously known as Sprints now have reviews added. This allows more opportunities for teams to incorporate improvements. Additionally, a Team Demo has been added in each iteration review. This provides more opportunity for transparency, sharing, and feedback.Development Team: The Development team was specifically identified at the team level in SAFe® 4.5. The development team is made up of three to nine people who can move an element of work from development through the test. This development team contains software developers, testers, and engineers, and does not include the Product Owner and Scrum Master. Each of those roles is shown separately at the team level in SAFe® 4.5.Scrum events: The list of scrum events are shown next to the ScrumXP icon and include Plan, Execute, Review, and Retro (for a retrospective.)Combined SAFe® Foundation Elements SAFe® 4.0 had the foundational elements of Core Values, Lean-Agile Mindset, SAFe® Principles, and Implementing SAFe® at a basic level.SAFe® 4.5 adds to the foundation elements by also including Lean-Agile Leaders, the Implementation Roadmap, and the support of the SPC in the successful implementation of SAFe®.Additional changes include: Communities of Practice: This was moved to the spanning palette to show support at all levels: team, program, large solution, and portfolio.Lean-Agile Leaders: This role is now included in the foundational level. Supportive leadership is critical to a successful SAFe® adoption.SAFe® Program Consultant: This role was added to the Foundational Layer. The SPC can play a key leadership role in a successful transition to Scaled Agile.Implementation Roadmap: The implementation roadmap replaces the basic implementation information in SAFe® 4.0. It provides more in-depth information on the elements to a successful enterprise transition to SAFe®.Benefits of upgrading to SAFe® 4.5With the addition of Lean Startup approaches, along with a deeper focus on DevOps and Continuous Delivery, teams will be situated to deliver quality and value to users more quickly.With improvements at the Portfolio level, teams get more guidance on Portfolio governance and other portfolio levels concerns, such as budgeting and compliance.  Reasons to Upgrade to SAFe® 4.5 Enterprises who’ve been using SAFe® 4.0 will find greater flexibility with the added levels in SAFe® 4.5. Smaller groups in the enterprise can use the team level, while groups working on more complex initiatives can create Agile Release Trains with many teams.Your teams can innovate faster by using the Lean Startup Approach. Work with end users to identify the Minimum Viable Product (MVP), then iterate as you get fast feedback and adjust. This also makes your customer more of a partner in development, resulting in better collaboration and a better end product.Get features and value to your user community faster with DevOps and the Continuous Delivery pipeline. Your teams can continuously hypothesize, build, measure, and learn to continuously release value. This also allows large organizations to innovate more quickly.Most Recent Changes in SAFe® series - SAFe® 4.6Because Scaled Agile continues to improve, new changes have been incorporated with SAFe® 4.6. with the addition of five core competencies that enable enterprises to respond to technology and market changes.Lean Portfolio Management: The information needed for how to use a Lean-Agile approach to portfolio strategy, funding, and governance.Business Solutions and Lean Systems: Optimizing activities to Implement large, complex initiatives using a Scaled Agile approach while still addressing the necessary activities such as designing, testing, deployment, and even retiring old solutions.DevOps and Release on Demand: The skills needed to release value as needed through a continuous delivery pipeline.Team and Technical Agility: The skills needed to establish successful teams who consistently deliver value and quality to meet customer needs.Lean-Agile Leadership: How leadership enables a successful agile transformation by supporting empowered teams in implementing agile practices. Leaders carry out the Agile principles and practices and ensure teams have the support they need to succeedSAFe® Agilist (SA) Certification exam: The SAFe® Agilist certification is for the change leaders in an organization to learn about the SAFe® practices to support change at all levels: team, program, and portfolio levels. These change agents can play a positive role in an enterprise transition to SAFe®.In order to become certified as a SAFe® Agilist (SA), you must first take the Leading SAFe® class and pass the SAFe® certification exam. To learn more about this, see this article on How To Pass Leading SAFe® 4.5 Exam.SAFe® Certification Exam: KnowledgeHut provides Leading SAFe® training in multiple locations. Check the site for locations and dates.SAFe® Agile Certification Cost: Check KnowledgeHut’s scheduled training offerings to see the course cost. Each course includes the opportunity to sit for the exam included in the cost.Scaled Agile Framework Certification Cost: There are multiple levels of SAFe® certification, including Scrum Master, Release Train Engineer, and Product Owner. Courses range in cost, but each includes the chance to sit for the corresponding SAFe® certification.SAFe® Classes: SAFe® classes are offered by various organizations. To see if KnowledgeHut is offering SAFe® Training near you, check the SAFe® training schedule on our website.TrainingKnowledgeHut provides multiple Scaled Agile courses to give both leaders and team members in your organization the information they need to for a successful transition to Scaled Agile. Check the site for the list of classes to find those that are right for your organization as you make the journey.All course fees cover examination costs for certification.SAFe® 4.5 Scrum Master with SSM Certification TrainingLearn the core competencies of implementing Agile across the enterprise, along with how to lead high-performing teams to deliver successful solutions. You’ll also learn how to implement DevOps practices. Completion of this course will prepare you for obtaining your SAFe® 4 Scrum Master certificate.SAFe® 4 Advanced Scrum Master (SASM)This two-day course teaches you to how to apply Scrum at the enterprise level and prepares you to lead high-performing teams in a Scaled Agile environment. At course completion, you’ll be prepared to manage interactions not only on your team but also across teams and with stakeholders. You’ll also be prepared to take the SAFe® Advanced Scrum Master exam.Leading SAFe®4.5 Training Course (SA)This two-day Leading SAFe® class prepares you to become a Certified SAFe® 4 Agilist, ready to lead the agile transformation in your enterprise.  By the end of this course, you’ll be able to take the SAFe® Agilist (SA) certification exam.SAFe® 4.5 for Teams (SP) This two-day course teaches Scrum fundamentals, principles tools, and processes. You’ll learn about software engineering practices needed to scale agile and deliver quality solutions in a Scaled Agile environment. Teams new to Scaled Agile will find value in going through this course. Attending the class prepares you for the certification exam to become a certified SAFe® 4 Practitioner (SP). DevOps Foundation Certification trainingThis course teaches you the DevOps framework, along with the practices to prepare you to apply the principles in your work environment. Completion of this course will prepare you also to take the DevOps Foundation exam for certification.
5175
A Glimpse Of The Major Leading SAFe® Versions

A Quick view of SAFe® Agile has gained popularit... Read More

How Start Ups Can Benefit From Cloud Computing?

From nebulous beginnings, the cloud has grown to a platform that has gained universal acceptance and is transforming businesses across industries. Companies that have adopted cloud technology have seen significant payoffs, with cloud based tools redefining their data storage, data sharing, marketing and project management capabilities. The easy availability of affordable cloud infrastructure has made it so easy to set up new businesses that the economy is all set for a start up boom which has its head, so to speak, in the cloud! With the advent of this new technology, complete newbie’s in the market are able to hold their own against established market players—by achieving an amazing quantum of work using skeleton manpower resources. Recently, a popular ad doing the rounds on TV showed a long haired youth conducting business from a cafe on his HP Pavilion laptop, where he is ridiculed by some well heeled middle aged businessmen on their coffee break. Back at their office, they find that this youngster is the new investor that their boss has been heaping accolades on. “Where’s your office?” one of them asks the young man…to be laughingly told that he carries his entire office in his laptop! And that, typically, is how the new-age start up business looks. We have heard many stories of how a clever idea has turned a tidy profit for a smart entrepreneur working out of his laptop. While cloud computing is pushing the boundaries of science and innovation into a new realm, it is also laying the foundation for a new wave of business start ups. New ventures in general suffer from a lack of infrastructure, manpower and funding…and all these three concerns are categorically addressed by the cloud. Moving to the cloud minimizes the need of huge capital investments to set up expensive infrastructure. For nascent entrepreneurs, physical hardware and server costs used to be formidable given the limited budgets at their disposal. Seed money was also required to hire office space, promote the business and hire workers. Today, thanks to cloud technology, getting a new business off the ground and running costs virtually nothing. Most of the resources and tools that new ventures need are available on the cloud at minimal costs, in fact quite often at zero costs, making this a powerful value proposition for small businesses. A cloud hosting provider such as AWS can enable you to go live immediately, and will even scale up to your requirement once your business expands. Small businesses can think and dream big with the cloud. When it comes to manpower resources, it takes just a handful of people to work wonders using the online resources that are at their disposal. If you have a brilliant idea and have a workable plan for execution, you can comfortably compete neck to neck with market leaders. The messaging sensation WhatsApp was started in 2009 by just two former Yahoo employees who leveraged the power of the internet – and this goes to show that clever use of technology can completely eliminate the need for a sizeable manpower pool. Start ups have always been more agile than their large scale counterparts, and the cloud helps them take this a step further. Resources can be scaled up or down in no time, whereas in traditional environments it would have taken many days, considerable planning and funds to add hardware and software. Cloud computing also helps improve collaboration across teams, often across geographies. Data sharing is instantaneous, and teams can work on a task together in real time regardless of their location. Powered by the cloud, small businesses operate with shoestring budgets and key players in different continents. All their accounting, client data, marketing and other business critical files can be stored online and are accessible from anywhere. These online tools can be accessed and utilised instantly, and underpin all the crucial processes on which these businesses thrive. Strategic financial decisions are made after garnering insights from cloud-based accounting software. E-invoicing helps settle bills in a fraction of the time of traditional billing systems, and client queries are answered quickly through cloud-based management systems—saving precious time and increasing customer satisfaction levels to an all-time high. Whether at home, on vacation or on the phone, businesses can oversee sales, replenish products and plan new sales strategies. That’s a whole new way of doing business, and seems to be very successful! An estimate by Cloudworks has put the anticipated cloud computing market at over $200 billion by the year 2018. As Jeff Weiner, CEO of LinkedIn, succinctly put it, the cloud “makes it easier and cheaper than ever for anyone anywhere to be an entrepreneur and to have access to all the best infrastructure of innovation.” With cloud technology rapidly levelling the playing field between nascent and established businesses, it is anybody’s guess as to just how many new start ups will burst into the scene in the next few years. Hoping that the blog has helped you gain a clear understanding of the importance of Cloud Computing.  To gain more knowledge on what cloud computing has to offer, take a look at other blogs as well as the AWS certifications that we have to offer or enrol yourself for the AWS Certification Training course by KnowledgeHut.  
How Start Ups Can Benefit From Cloud Computing?

From nebulous beginnings, the cloud has grown to a... Read More

Business Transformation through Enterprise Cloud Computing

The Cloud Best Practices Network is an industry solutions groups and best practices catalogue of how-to information for Cloud Computing. While we cover all aspects of the technology our primary goal is to explain the enabling relationship between this new IT trend and business transformation, where our materials include: Core Competencies – The mix of new skills and technologies required to successfully implement new Cloud-based IT applications. Reference Documents – The core articles that define what Cloud Computing is and what the best practices are for implementation, predominately referring to the NIST schedule of information. Case studies – Best practices derived from analysis of pioneer adopters, such as the State of Michigan and their ‘MiCloud‘ framework . Read this article ‘Make MiCloud Your Cloud‘ as an introduction to the Cloud & business transformation capability. e-Guides – These package up collections of best practice resources directed towards a particular topic or industry. For example our GovCloud.info site specializes in Cloud Computing for the public sector. White papers – Educational documents from vendors and other experts, such as the IT Value mapping paper from VMware. Core competencies The mix of new skills and technologies required to successfully implement new Cloud-based IT applications, and also the new capabilities that these platforms make possible: Virtualization Cloud Identity and Security – Cloud Privacy Cloud 2.0 Cloud Configuration Management Cloud Migration Management DevOps Cloud BCP ITaaS Procurement Cloud Identity and Security Cloud Identity and Security best practices (CloudIDSec) provides a comprehensive framework for ensuring the safe and compliant use of Cloud systems. This is achieved through combining a focus on the core references for Cloud Security, the Cloud Security Alliance, with those of Cloud Identity best practices: IDaaS – Identity Management 2.0 Federated Identity Ecosystems Cloud Privacy A common critcal focus area for Cloud computing is data privacy, particularly with regards to the international aspects of Cloud hosting. Cloud Privacy refers to the combination of technologies and legal frameworks to ensure privacy of personal information held in Cloud systems, and a ‘Cloud Privacy-by-Design’ process can then be used to identify the local legislated privacy requirements of information. Tools for designing these types of privacy controls have been developed by global privacy experts, such as Ann Cavoukian, the current Privacy Commissioner for Ontario, who provides tools to design and build these federated privacy systems. The Privacy by Design Cloud Computing Architecture (26-page PDF) document provides a base reference for how to combine traditional PIAs (Privacy Impact Assessments) with Cloud Computing. As this Privacy Framework presentation then explains these regulatory mechanisms that Kantara enables can then provide the foundations for securing the information in a manner that encompasses all the legacy, privacy and technical requirements needed to ensure it is suitable for e-Government scenarios. This then enables it to achieve compliance with the Cloud Privacy recommendations put forward by global privacy experts, such as Ann Cavoukian, the current Privacy Commissioner for Ontario, who stipulates a range of ‘Cloud Privacy By Design‘ best practices Cloud 2.0 Cloud is as much a business model as it is a technology, and this model is best described through the term ‘Cloud 2.0′. As the saying goes a picture tells a thousand words, and as described by this one Cloud 2.0 represents the intersection between social media, Cloud computing and Crowdsourcing. The Social Cloud In short it marries the emergent new online world of Twitter, Linkedin et al, and the technologies that are powering them, with the traditional, back-end world of mainframe systems, mini-computers and all other shapes and sizes of legacy data-centre. “Socializing” these applications means moving them ‘into the Cloud’, in the sense of connecting them into this social data world, as much as it does means virtualizing the application to run on new hardware. This a simple but really powerful mix, that can act as a catalyst for an exciting new level of business process capability. It can provide a platform for modernizing business processes in a significant and highly innovative manner, a breath of fresh air that many government agency programs are crying out for. Government agencies operate many older technology platforms for many of their services, making it difficult to amend them for new ways of working and in particular connecting them to the web for self-service options. Crowdsourcing Social media encourages better collaboration between users and information, and tools for open data and back-end legacy integrations can pull the transactional systems informtion needed to make this functional and valuable. Crowdsourcing is: a distributed problem-solving and production process that involves outsourcing tasks to a network of people, also known as the crowd. Although not a component of the technologies of Cloud Computing, Crowdsourcing is a fundamental concept inherent to the success of the Cloud 2.0 model. The commercial success of migration to Cloud Computing will be amplified when there is a strong focus on the new Web 2.0 type business models that the technology is ideal for enabling. Case study – Peer to Patent One such example is the Whitehouse project the Peer to the Patent portal, a headline example of Open Government, led by one its keynote experts Beth Noveck. This project illustrates the huge potential for business transformation that Cloud 2.0 offers. It’s not just about migrating data-center apps into a Cloud provider, connecting an existing IT system to a web interface or just publishing Open Data reporting data online, but rather utilizing the nature of the web to entirely re-invent the core process itself. It’s about moving the process into the Cloud. In this 40 page Harvard white paper Beth describes how the US Patent Office was building up a huge backlog of over one million patent applications due to a ‘closed’ approach where only staff from the USPTO could review, contribute and decide upon applications. To address this bottleneck she migrated the process to an online, Open version where contributors from across multiple organizations could help move an application through the process via open participation web site features. Peer to Patent is a headline example of the power of Open Government, because it demonstrates its about far more than simply publishing reporting information online in an open manner, so that they public can inspect data like procurement spending numbers. Rather it’s about changing the core decision-making processes entirely, reinventing how Government itself works from the inside out, reinventing it from a centralized hierarchical monolith to an agile, distributed peer to peer network. In essence it transforms the process from ‘closed’ to ‘open’, in terms of who and how others can participate, utilizing the best practice of ‘Open Innovation‘ to break the gridlock that had occured due the constraints caused by private, traditional ways of working. Open Grantmaking – Sharing Cloud Best Practices Beth has subsequently advised further on how these principles can be applied in general across Government. For example in this article on her own blog she describes ‘Open Grantmaking‘ – How the Peer To Patent crowdsourcing model might be applied to the workflows for government grant applications. She touches on what is the important factor about these new models, their ability to accelerate continual improvement within organizations through repeatedly sharing and refining best practices: “In practice, this means that if a community college wins a grant to create a videogame to teach how to install solar panels, everyone will have the benefit of that knowledge. They will be able to play the game for free. In addition, anyone can translate it into Spanish or Russian or use it as the basis to create a new game to teach how to do a home energy retrofit.” Beth describes how Open Grantmaking might be utilized to improve community investing in another blog, describing how OG would enable more transparency and related improvements. Cloud 2.0 As the underlying technology Cloud 2.0 caters for both the hosting of the software and also the social media 2.0 features that enable the cross-enterprise collaboration that Beth describes. Cloud Configuration Management CCM is the best practice for change and configuration management within Cloud environments, illustrated through vendors such as Evolven. Problem Statement One of the key goals and perceived benefits of Cloud computing is a simplified IT environment, a reduction of complexity through virtualizing applications into a single overall environment. However complexity actually increases.  Virtual Machines (VMs) encapsulate application and infrastructure configurations, they package up a combination of applications and their settings, obscuring this data from traditional configuration management tools. Furthermore the ease of self-service creation of VMs results in their widespread proliferation, and so actually the adoption of Cloud technologies creates a need for a new, extra dimension of systems management. This is called CCM, and incorporates: Release & Incident Management The increased complexity therefore increases the difficulties in trouble-shooting technical problems, and thus requires an updated set of tools and also updates to best practices like the use of ITIL procedures. ‘Release into Production’ is a particularly sensitive process within software teams, as major upgrades and patches are transitioned from test to live environments. Any number of configuration-related errors could cause the move to fail, and so CCM software delivers the core competency of being better able to respond quicker to identify and resolve these issues, reducing the MTTR significantly. DevOps DevOps is a set of principles, methods and practices for communication, collaboration and integration between software development and IT operations. Through the implementation of a shared Lean adoption program and QMS (Quality Management System) the two groups can better work together to minimize downtimes while improving the speed and quality of software development. It’s therefore directly linked to Business Agility. The higher the value of speed and quality = a faster ability to react to market changes, deploy new products and processes and in general adapt the organization, achieved through increasing the frequency of ‘Release Events’: It’s therefore directly linked to Business Agility. The higher the value of speed and quality = a faster ability to react to market changes, deploy new products and processes and in general adapt the organization, achieved through increasing the frequency of ‘Release Events’: ITaaS Procurement The fundamental shift that Cloud Computing represents is illustrated in one key implementation area:   Procurement. Moving to Cloud services means changing from a financial model for technology where you buy your own hardware and software, and pay for it up front, to an approach where instead you access it as a rental, utility service where you “PAYG – Pay As You Go”. To encompass all the different ‘as a Service’ models this is known at an overall level as ‘ITaaS’ – IT as a Service. Any type of IT can be virtualized and delivered via this Service model. Towards the end, I hope that you have gained a clear understanding of How Business Transforms Through Enterprise Cloud Computing. If this article has helped you clear your fundamentals and if you wish to learn more about Cloud computing by getting certified, then you can undertake the AWS certification course offered by KnowledgeHut.
Business Transformation through Enterprise Cloud C...

The Cloud Best Practices Network is an industry ... Read More