d43ec162f0
Unforunately, It came to my knowledge that this repository promoted a phenomenon where DevOps interviews became a trivia game where people think it's normal to throw 20 random short questions like "what is fork()" or "which tools would you use for each of the following areas?" and this was not my intention. To explcitly state this repository doesn't represents real DevOps interview questions I've decided to rename it.
23 lines
1.0 KiB
Markdown
23 lines
1.0 KiB
Markdown
# Elasticsearch, Kibana and AWS - Solution
|
|
|
|
This one out of many possible solutions. This solution is relying heavily on AWS.
|
|
|
|
* Create a VPC with subnet so we can place Elasticsearch node(s) in internal environment only.
|
|
If required, we will also setup NAT for public access.
|
|
|
|
* Create an IAM role for the access to the cluster. Also, create a separate role for admin access.
|
|
|
|
* To provision the solution quickly, we will use the elasticsearch service directly from AWS for production deployment.
|
|
This way we also cover multiple AZs. As for authentication, we either use Amazon cognito or the organization LDAP server.
|
|
|
|
* To transfer data, we will have to install logstash agent on the instances. The agent will be responsible
|
|
for pushing the data to elasticsearch.
|
|
|
|
* For monitoring we will use:
|
|
|
|
* Cloud watch to monitor cluster resource utilization
|
|
* Cloud metrics dashboard
|
|
|
|
* If access required from multiple regions we will transfer all the data to S3 which will allow us to view the data
|
|
from different regions and consolidate it in one dashboard
|