Hello, community!
I’m experiencing an issue while setting up an automation in Make.com for image analysis using the GPT module. My automation is structured as follows:
- First Module — Telegram → Watch Updates:
- Receives a user message, which may include an attached image in JPG format.
- Second Module — Telegram → Download a File:
- Downloads the attached image.
- Third Module — OpenAI GPT → Analyze Image:
- Sends the downloaded image to GPT for analysis.
Automation Goal: The user sends an image along with a text query (e.g., “What is shown in the photo?” or “Translate the text in the instruction”). GPT should analyze the content of the image and provide a relevant response.
Problem:
- The GPT → Analyze Image module provides incorrect and irrelevant responses:
- For example, instead of accurately describing the image, GPT gives an incorrect interpretation.
- When the image contains text, GPT fails to extract it correctly.
- However, when the same image is uploaded to GPT via the web interface, it is processed correctly, and the responses are accurate.
My Actions:
- I confirmed that the file is downloaded and passed correctly to the GPT module (the JPG format is preserved).
- I tested the same images via the web interface, where GPT produced correct results.
Questions:
- Why does the GPT module in Make.com process the image incorrectly, while the web interface provides accurate results?
- What steps can I take to improve the accuracy of image analysis in Make.com?
- Are there specific recommendations for passing images or parameters that could affect the GPT module’s performance?
Thank you in advance for your assistance!