Skip to content
This repository was archived by the owner on Apr 8, 2025. It is now read-only.
This repository was archived by the owner on Apr 8, 2025. It is now read-only.

Show how the LLM is involved in the MCP workflow #53

Open
@hgmuc

Description

@hgmuc

I would appreciate, if the high-level architecture diagrams would also show the data flow (requests, response, message payloads from the context) to and from the LLM. From the weather-server example with Claude Desktop it seems to me that Claude Desktop calls the LLM to create the tool call with the correct arguments. But I can't see how the returned forecast data will be formatted in a user-friendly way. Is Claude Desktop passing the forecast data to the LLM to get a nicely formatted answer or is Claude Desktop able to show this data in a nicely formatted way (e.g. table) out of the box?

  • Which component (host, server) calls the LLM and when?
  • Or is it up to developer to choose if the host or the server calls the LLM when it seems to be best from the workflow perspective?
  • Are there any recommendations or best practises for getting the LLM involved?

Finally, adding the LLM to high-level diagrams would also help to understand potential risks when using MCP with sensitive data. Claude Desktop obviously uses Claude LLMs via the internet. But with sensitive data one might consider using a local LLM, which would maybe called by the MCP server.

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions