Data Sources
RunLLM can learn from a wide variety of data sources, including your documentation, customer conversations, and codebases. Each data connector is backed by a custom-built data pipeline that reads, classifies, and annotates the information as its ingested into your assistant's knowledge base.
Configuring data sources is easy. Navigate to your the config tab of your assistant page, click Add New next to Data, and select the data source you would like to add. Fill out the form on the next page, give RunLLM any necessary permissions, and you're done!
Once you add a data source, you can track its ingestion progress from the RunLLM dashboard. When you register the source, you should see a new entry under the data table on the config tab. Clicking on the entry will show you how much data RunLLM's ingested and what work is left.
RunLLM supports the following integrations:
Knowledge
- Confluence documents
- Cortex
- Discord channels
- Discourse forums
- File uploads
- Freshdesk
- GitHub Issues, Discussions, and PRs
- GitHub Repos
- Google Drive
- Guru cards
- Intercom tickets
- Jira tickets
- Linear
- Notion documents
- Plain
- Salesforce
- Sharepoint documents
- Slack channels
- Web crawler
- YouTube videos
- Zendesk tickets
SRE Tools
If you find that something you'd like to see is missing, please don't hesitate to reach out. Here are some of the data sources we're currently working on adding:
- Salesforce tickets
- Slab documents
Data Source Configuration
There are a handful of parameters that are shared across every data source:
- Name: Every data source requires a name. This should hopefully be self-explanatory.
- Update schedule: By default, every data source will be ingested monthly. You can set up update schedule under the advanced configuration for each data source. A daily update schedule will update every midnight, a weekly schedule on Sunday at midnight, and a monthly schedule on the 1st at midnight. If you would like to specify a Cron string, you can use a
Customschedule. - Data source group: By default, this should be left blank. This field can be used if there are priorities amongst different categories of data (e.g., prioritize documentation over Slack messages). **Please contact our team if you'd like to set this up.
Sometimes, data needs to be updated between scheduled ingestion jobs. To update your data source, go to the RunLLM dashboard, click on the data source you'd like to update, and click the Run Now button. Within a few minutes, your data source should be updated!