-
Notifications
You must be signed in to change notification settings - Fork 384
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Unexpected LlamaParse Behavior in v0.6.1 – Extracting Raw OCR Instead of Analyzing Page Content #621
Comments
Hmmm. The old params should still work, so you can still set I do see the same issue when trying the new params myself though |
Thank you for your response! I appreciate your time and effort in investigating this issue. The issue still persists. I just tested it again, and the bug is still present in versions 0.6.0 and 0.6.1. Job ID: For now, I will continue using the parsing_instruction parameter as a workaround. However, I am still looking forward to an update regarding this issue, as I would prefer to use complemental_formatting_instruction, content_guideline_instruction, and formatting_instruction. If these parameters start working correctly again, would it be advisable to lock my project to version 0.6.0 or 0.6.1 to continue using them (complemental_formatting_instruction, content_guideline_instruction, and formatting_instruction)? Or would migrating to the newly introduced parameters be the better approach? However, I also noticed that the documentation and GUI now mention additional parameters (user_prompt, system_prompt_append, and system_prompt). Would it be advisable to migrate my code to these newer parameters instead? Or will the previously introduced parameters be fully supported again in future versions? Thanks again for your assistance! Looking forward to your insights. |
Tested this again today in the sandbox, and although the prompt is visible in the job details, the output still seems like raw OCR and has not followed the instruction. As you say parseing_instruction still working via the API at the moment |
It's day 4 of this issue and even though parsing instruction may work with http API call, it isn't working with the python library. I tried the GUI today, and instructions are not working there either. If anyone has any suggestions for an alternative API, do suggest as my code for a client is breaking in production |
Facing the same issues here, any workaround for this, would help unblock me in my task |
I've checked the current status of LlamaCloud on llamaindex.statuspage.io, and there is an ongoing issue reported as "Degraded performance on LlamaCloud". The latest update from the team indicates that they are still investigating the issue. Currently, LlamaCloud Public API and LlamaParse are experiencing a Major Outage, which has been ongoing for about 55 minutes. Before this full outage, LlamaCloud Public API was already in Partial Outage for approximately 2 hours and 40 minutes. Given this, it seems best to wait for the service to stabilize before testing further. Once LlamaCloud is fully operational again, it will be important to verify whether the issues with LlamaParse's structured output persist. The discussions and shared test results have been useful in understanding how this issue is impacting different use cases. The ongoing updates help provide clarity on the situation. I appreciate the team's efforts in investigating this, and I’ll keep an eye on the status updates. If anyone has observed differences in service behavior across different regions, it could be helpful to share those findings. Looking forward to updates. |
I have tried using the premium mode just in case and it is working better obviously. Even the parsing time isn't delayed on this mode. I also saw a new PR related to user_prompt and related fields, and I think these changes and merges definitely have something to do with the outage in the first place. I am really hopeful that the problem may resolve soon now that it is getting attention from the providers. |
It seems that approximately 1.5 hours ago, the team provided an update stating that they applied a patch to restore some functionality to their services. However, there are still noticeable performance issues affecting the system. From my recent testing, the GUI interface appears to be functioning properly on my end. However, when using the API, I encountered two specific issues:
From my observations, "NO CONTENT HERE" seems to occur specifically when calling third-party LLMs through the API. It looks like the API is unable to interact with the selected LLM properly, leading to this response. I was just running some troubleshooting tests to check the current state of the system and wanted to report my findings based on the latest conditions. My team is currently waiting for a resolution, but in the meantime, I’ve had to resort to alternative solutions to keep production running. We’re hopeful that the service will return to its expected stability as soon as possible, not just for me and my team, but also for the other users who have reported similar issues. Looking forward to seeing everything back up and running smoothly soon. |
@mehwishh247 . The PR#622, which introduces the new parameters (user_prompt, system_prompt, and system_prompt_append), is still open and has not yet been merged. This means these parameters are not available for use in the API at the moment. We'll need to wait for further updates. I also commented on PR#622 asking if there's any update or an estimated timeline for when these parameters will be available. Let's stay tuned for any progress. |
Hi everyone, Following my last comment, PR#622 has been merged yesterday, making the new parameters (user_prompt, system_prompt, and system_prompt_append) available for the API. I’d like to thank the maintainer responsible for the merge—this update is greatly appreciated, as many users were eagerly waiting for these parameters. Today, I have updated all my projects codes to version 0.6.2, ensuring full compatibility with the new changes. I’m now waiting for my team to validate that everything is working as expected with LlamaParse v0.6.2. Once my team confirms that the functionality is stable, I will proceed with closing this issue. Thanks again for the support. |
Hey @hitoshyamamoto, |
Describe the bug
I frequently use the LlamaParse method to obtain responses based on the content of the page image I am working on. However, after upgrading to v0.6.1, which introduced new parameters (content_guideline_instruction, formatting_instruction, complemental_formatting_instruction), I noticed an issue.
The responses are no longer as expected. Instead of analyzing the page and generating an appropriate response, LlamaParse is behaving like a simple OCR, extracting raw text from the image without considering the intended formatting or content structure.
LlamaParse should analyze the page image and generate a structured response based on the provided instructions, rather than just extracting plain text.
Job ID
7f9033bf-723b-475b-bcb8-7100074eebc2
Client:
Please remove untested options:
Additional context
The text was updated successfully, but these errors were encountered: