- Kibana Guide: other versions:
- What is Kibana?
- What’s new in 8.16
- Kibana concepts
- Quick start
- Set up
- Install Kibana
- Configure Kibana
- Alerting and action settings
- APM settings
- Banners settings
- Cases settings
- Enterprise Search settings
- Fleet settings
- i18n settings
- Logging settings
- Logs settings
- Metrics settings
- Monitoring settings
- Reporting settings
- Search sessions settings
- Secure settings
- Security settings
- Spaces settings
- Task Manager settings
- Telemetry settings
- URL drilldown settings
- Start and stop Kibana
- Access Kibana
- Securing access to Kibana
- Add data
- Upgrade Kibana
- Configure security
- Configure reporting
- Configure logging
- Configure monitoring
- Command line tools
- Production considerations
- Discover
- Dashboards
- Canvas
- Maps
- Build a map to compare metrics by country or region
- Track, visualize, and alert on assets in real time
- Map custom regions with reverse geocoding
- Heat map layer
- Tile layer
- Vector layer
- Plot big data
- Search geographic data
- Configure map settings
- Connect to Elastic Maps Service
- Import geospatial data
- Troubleshoot
- Reporting and sharing
- Machine learning
- Graph
- Alerting
- Observability
- Search
- Security
- Dev Tools
- Fleet
- Osquery
- Stack Monitoring
- Stack Management
- Cases
- Connectors
- Amazon Bedrock
- Cases
- CrowdStrike
- D3 Security
- Google Gemini
- IBM Resilient
- Index
- Jira
- Microsoft Teams
- Observability AI Assistant
- OpenAI
- Opsgenie
- PagerDuty
- SentinelOne
- Server log
- ServiceNow ITSM
- ServiceNow SecOps
- ServiceNow ITOM
- Swimlane
- Slack
- TheHive
- Tines
- Torq
- Webhook
- Webhook - Case Management
- xMatters
- Preconfigured connectors
- License Management
- Maintenance windows
- Manage data views
- Numeral Formatting
- Rollup Jobs
- Manage saved objects
- Security
- Spaces
- Advanced Settings
- Tags
- Upgrade Assistant
- Watcher
- REST API
- Get features API
- Kibana spaces APIs
- Kibana role management APIs
- User session management APIs
- Saved objects APIs
- Data views API
- Index patterns APIs
- Alerting APIs
- Action and connector APIs
- Cases APIs
- Import and export dashboard APIs
- Logstash configuration management APIs
- Machine learning APIs
- Osquery manager API
- Short URLs APIs
- Get Task Manager health
- Upgrade assistant APIs
- Synthetics APIs
- Uptime APIs
- Kibana plugins
- Troubleshooting
- Accessibility
- Release notes
- Upgrade notes
- Kibana 8.16.4
- Kibana 8.16.3
- Kibana 8.16.2
- Kibana 8.16.1
- Kibana 8.16.0
- Kibana 8.15.5
- Kibana 8.15.4
- Kibana 8.15.3
- Kibana 8.15.2
- Kibana 8.15.1
- Kibana 8.15.0
- Kibana 8.14.3
- Kibana 8.14.2
- Kibana 8.14.1
- Kibana 8.14.0
- Kibana 8.13.4
- Kibana 8.13.3
- Kibana 8.13.2
- Kibana 8.13.1
- Kibana 8.13.0
- Kibana 8.12.2
- Kibana 8.12.1
- Kibana 8.12.0
- Kibana 8.11.4
- Kibana 8.11.3
- Kibana 8.11.2
- Kibana 8.11.1
- Kibana 8.11.0
- Kibana 8.10.4
- Kibana 8.10.3
- Kibana 8.10.2
- Kibana 8.10.1
- Kibana 8.10.0
- Kibana 8.9.2
- Kibana 8.9.1
- Kibana 8.9.0
- Kibana 8.8.2
- Kibana 8.8.1
- Kibana 8.8.0
- Kibana 8.7.1
- Kibana 8.7.0
- Kibana 8.6.1
- Kibana 8.6.0
- Kibana 8.5.2
- Kibana 8.5.1
- Kibana 8.5.0
- Kibana 8.4.3
- Kibana 8.4.2
- Kibana 8.4.1
- Kibana 8.4.0
- Kibana 8.3.3
- Kibana 8.3.2
- Kibana 8.3.1
- Kibana 8.3.0
- Kibana 8.2.3
- Kibana 8.2.2
- Kibana 8.2.1
- Kibana 8.2.0
- Kibana 8.1.3
- Kibana 8.1.2
- Kibana 8.1.1
- Kibana 8.1.0
- Kibana 8.0.0
- Kibana 8.0.0-rc2
- Kibana 8.0.0-rc1
- Kibana 8.0.0-beta1
- Kibana 8.0.0-alpha2
- Kibana 8.0.0-alpha1
- Developer guide
Playground
editPlayground
editThis functionality is in technical preview and may be changed or removed in a future release. Elastic will work to fix any issues, but features in technical preview are not subject to the support SLA of official GA features.
Use Playground to combine your Elasticsearch data with the power of large language models (LLMs) for retrieval augmented generation (RAG). The chat interface translates your natural language questions into Elasticsearch queries, retrieves the most relevant results from your Elasticsearch documents, and passes those documents to the LLM to generate tailored responses.
Once you start chatting, use the UI to view and modify the Elasticsearch queries that search your data. You can also view the underlying Python code that powers the chat interface, and download this code to integrate into your own application.
Learn how to get started on this page. Refer to the following for more advanced topics:
How Playground works
editHere’s a simpified overview of how Playground works:
- User creates a connection to LLM provider
- User selects a model to use for generating responses
-
User define the model’s behavior and tone with initial instructions
- Example: "You are a friendly assistant for question-answering tasks. Keep responses as clear and concise as possible."
- User selects Elasticsearch indices to search
- User enters a question in the chat interface
-
Playground autogenerates an Elasticsearch query to retrieve relevant documents
- User can view and modify underlying Elasticsearch query in the UI
-
Playground auto-selects relevant fields from retrieved documents to pass to the LLM
- User can edit fields targeted
-
Playground passes filtered documents to the LLM
- The LLM generates a response based on the original query, initial instructions, chat history, and Elasticsearch context
-
User can view the Python code that powers the chat interface
- User can also Download the code to integrate into application
Availability and prerequisites
editFor Elastic Cloud and self-managed deployments Playground is available in the Search space in Kibana, under Content > Playground.
For Elastic Serverless, Playground is available in your Elasticsearch project UI.
To use Playground, you’ll need the following:
- An Elastic v8.14.0+ deployment or Elasticsearch Serverless project. (Start a free trial).
-
At least one Elasticsearch index with documents to search.
- See ingest data if you’d like to ingest sample data.
-
An account with a supported LLM provider. Playground supports the following:
Provider Models Notes Amazon Bedrock
- Anthropic: Claude 3.5 Sonnet
- Anthropic: Claude 3 Haiku
OpenAI
- GPT-3 turbo
- GPT-4 turbo
- GPT-4 omni
Azure OpenAI
- GPT-3 turbo
- GPT-4 turbo
Buffers responses in large chunks
Google
- Google Gemini 1.5 Pro
- Google Gemini 1.5 Flash
Getting started
edit
Connect to LLM provider
editTo get started with Playground, you need to create a connector for your LLM provider. You can also connect to locally hosted LLMs which are compatible with the OpenAI API, by using the OpenAI connector.
To connect to an LLM provider, follow these steps on the Playground landing page:
- Under Connect to an LLM, click Create connector.
- Select your LLM provider.
- Name your connector.
- Select a URL endpoint (or use the default).
- Enter access credentials for your LLM provider. (If you’re running a locally hosted LLM using the OpenAI connector, you must input a value in the API key form, but the specific value doesn’t matter.)
If you need to update a connector, or add a new one, click the 🔧 Manage button beside Model settings.
Ingest data (optional)
editYou can skip this step if you already have data in one or more Elasticsearch indices.
There are many options for ingesting data into Elasticsearch, including:
- The Elastic crawler for web content (NOTE: Not yet available in Serverless)
- Elastic connectors for data synced from third-party sources
-
The Elasticsearch Bulk API for JSON documents
Expand for example
To add a few documents to an index called
books
run the following in Dev Tools Console:POST /_bulk { "index" : { "_index" : "books" } } {"name": "Snow Crash", "author": "Neal Stephenson", "release_date": "1992-06-01", "page_count": 470} { "index" : { "_index" : "books" } } {"name": "Revelation Space", "author": "Alastair Reynolds", "release_date": "2000-03-15", "page_count": 585} { "index" : { "_index" : "books" } } {"name": "1984", "author": "George Orwell", "release_date": "1985-06-01", "page_count": 328} { "index" : { "_index" : "books" } } {"name": "Fahrenheit 451", "author": "Ray Bradbury", "release_date": "1953-10-15", "page_count": 227} { "index" : { "_index" : "books" } } {"name": "Brave New World", "author": "Aldous Huxley", "release_date": "1932-06-01", "page_count": 268} { "index" : { "_index" : "books" } } {"name": "The Handmaids Tale", "author": "Margaret Atwood", "release_date": "1985-06-01", "page_count": 311}
We’ve also provided some Jupyter notebooks to easily ingest sample data into Elasticsearch. Find these in the elasticsearch-labs repository. These notebooks use the official Elasticsearch Python client.
Select Elasticsearch indices
editOnce you’ve connected to your LLM provider, it’s time to choose the data you want to search.
- Click Add data sources.
- Select one or more Elasticsearch indices.
- Click Save and continue to launch the chat interface.
You can always add or remove indices later by selecting the Data button from the main Playground UI.

Chat and query modes
editSince 8.15.0 (and earlier for Elasticsearch Serverless), the main Playground UI has two modes:
- Chat mode: The default mode, where you can chat with your data via the LLM.
- Query mode: View and modify the Elasticsearch query generated by the chat interface.
The chat mode is selected when you first set up your Playground instance.

To switch to query mode, select Query from the main UI.

Learn more about the underlying Elasticsearch queries used to search your data in View and modify queries
Set up the chat interface
editYou can start chatting with your data immediately, but you might want to tweak some defaults first.
You can adjust the following under Model settings:
- Model. The model used for generating responses.
- Instructions. Also known as the system prompt, these initial instructions and guidelines define the behavior of the model throughout the conversation. Be clear and specific for best results.
- Include citations. A toggle to include citations from the relevant Elasticsearch documents in responses.
Playground also uses another LLM under the hood, to encode all previous questions and responses, and make them available to the main model. This ensures the model has "conversational memory".
Under Indices, you can edit which Elasticsearch indices will be searched. This will affect the underlying Elasticsearch query.
Click ✨ Regenerate to resend the last query to the model for a fresh response.
Click ⟳ Clear chat to clear chat history and start a new conversation.
View and download Python code
editUse the View code button to see the Python code that powers the chat interface. You can integrate it into your own application, modifying as needed. We currently support two implementation options:
- Elasticsearch Python Client + LLM provider
- LangChain + LLM provider

Next steps
editOnce you’ve got Playground up and running, and you’ve tested out the chat interface, you might want to explore some more advanced topics:
On this page