# Third-party Integrations (118+)

OpenHuman ships with backend-proxied access to **118+ third-party services**. Connecting any of them is a one-click OAuth flow inside the app, there are no API keys to wire by hand, and no plugin marketplace to navigate.

(Under the hood, the connector layer is powered by [Composio](https://composio.dev). You will not need to think about it.)

Once a service is connected, it shows up in four places at once:

1. As an **agent tool**, the model can call it directly.
2. As a **memory source**, [auto-fetch](/openhuman/features/obsidian-wiki/auto-fetch.md) syncs it into the [Memory Tree](/openhuman/features/obsidian-wiki/memory-tree.md) every twenty minutes.
3. As a **profile signal**, your activity across services feeds your personalization.
4. As a **trigger source**, live events (a new email, a new charge, an inbound DM) flow into the [Triggers](/openhuman/features/integrations/triggers.md) pipeline and can fire off agent actions automatically.

## Some of what's in the catalog

The catalog spans productivity, business, social, messaging and Google. A non-exhaustive sample:

| Category                | Examples                                             |
| ----------------------- | ---------------------------------------------------- |
| **Email & calendar**    | Gmail, Outlook, Google Calendar, Apple Calendar      |
| **Docs & storage**      | Google Docs, Google Drive, Notion, Dropbox, Airtable |
| **Code & dev**          | GitHub, Linear, Jira, Figma                          |
| **Comms**               | Slack, Discord, Microsoft Teams, Telegram, WhatsApp  |
| **CRM & sales**         | Salesforce, HubSpot                                  |
| **Commerce & payments** | Stripe, Shopify                                      |
| **Project management**  | Asana, Trello                                        |
| **Social**              | Twitter / X, Spotify, YouTube                        |

## Native vs proxied

Some services have **native providers**. Rust modules that know how to ingest the service into the Memory Tree directly (e.g. Gmail's native ingest path). Others are exposed as **proxied tools** only: the agent can call them, but there's no automatic ingest yet. New native providers are added as features land.

## How connections work

Click **Connect** on any integration. A browser window opens for OAuth. Once you sign in, the connection becomes active and OpenHuman starts syncing it through [auto-fetch](/openhuman/features/obsidian-wiki/auto-fetch.md) on the next 20-minute tick.

Each integration shows its current status:

* **Not connected**. integration has not been set up.
* **Connected**. integration is active and being synced.
* **Manage**. active integration with options to reconfigure or disconnect.

You can revoke any connection at any time from the Skills tab.

## Messaging channels

Three integrations are special. OpenHuman uses them to *talk back* to you, not just read from them:

* **Telegram**. the primary messaging channel. Two-way: send and receive messages, manage chats, search history, create groups, 80+ actions on your behalf. All actions run through your own encrypted credentials.
* **Discord**. send and receive messages via Discord. Connect your account to receive OpenHuman messages there.
* **Web**. a browser-based chat interface within the desktop app. Messages stay entirely local.

Set your default under **Settings → Automation & Channels → Messaging Channels**. The active route status shows which channel is currently in use. Telegram offers two credential modes: connect via OpenHuman (one-click, encrypted) or provide your own credentials for maximum control.

## Skills

Beyond third-party services, OpenHuman has **skills**, small sandboxed modules that run inside the app, fetch external data, run on a schedule, transform information, and respond to events. Each runs with enforced resource limits. Skills install from the Skills tab and integrate with the same Memory Tree as everything else.

## Native voice and tools

Two capabilities ship native rather than as integrations because they're load-bearing for the desktop experience:

* [**Voice**](/openhuman/features/native-tools/voice.md). STT in, TTS out, plus a live Google Meet agent that joins meetings, transcribes them into your Memory Tree, and can speak back into the call.
* [**Native tools**](/openhuman/features/native-tools.md). built-in web search, web-fetch scraper, and a full filesystem/git/lint/test/grep coder toolset that the agent uses out of the box.

## Privacy boundary

OpenHuman's core never calls any third-party API directly. All requests go through the OpenHuman backend, which handles OAuth tokens and rate limiting. Your tokens never sit on disk in plaintext on your machine, and the agent only sees the *results* of tool calls, not the credentials.

See [Privacy & Security](/openhuman/features/privacy-and-security.md) for the full boundary.

## See also

* [Triggers](/openhuman/features/integrations/triggers.md), live events from connected integrations and how they fire agent actions.
* [Auto-fetch from Integrations](/openhuman/features/obsidian-wiki/auto-fetch.md)
* [Memory Tree](/openhuman/features/obsidian-wiki/memory-tree.md)


---

# Agent Instructions: Querying This Documentation

If you need additional information that is not directly available in this page, you can query the documentation dynamically by asking a question.

Perform an HTTP GET request on the current page URL with the `ask` query parameter:

```
GET https://tinyhumans.gitbook.io/openhuman/features/integrations.md?ask=<question>
```

The question should be specific, self-contained, and written in natural language.
The response will contain a direct answer to the question and relevant excerpts and sources from the documentation.

Use this mechanism when the answer is not explicitly present in the current page, you need clarification or additional context, or you want to retrieve related documentation sections.
