Data Story
I’ve made every effort to address the requests from The Treasury and The Public Services Commission. Based on my understanding, their main goal is to use AI agents to help create, edit, and review documents following Plain Language guidelines outlined in their style manual.
The challenge with government documents is that they’re subject to strict data privacy laws. Using cloud solutions would almost certainly cause issues since the data is typically processed overseas, making it unlikely to pass government scrutiny. Additionally, there’s the cybersecurity risk of protecting a large language model from prompt-hacking. To address these issues, I needed a solution that could run locally, on the government’s internal network, using affordable technology. It also had to be secure from prompt-hackers and flexible enough to work with multiple style guides, enhancing the document review process with AI agents.
This is where Code Blue, the document review multi-tool, comes into play.
Code Blue is a Python application that accepts style guides as user input. Once the style guides are uploaded, LM Studio processes the documents by creating embeddings (or chunks) and pulls together a unified list of rules from all the style guides. The LLAMA 3 language model then uses these rules to review a document uploaded by the user. The output is a text file that includes recommendations, noting specific paragraphs, page numbers, and suggested improvements.
Project Guidelines and How I Addressed These:
Competition Request:
"We are looking for creative applications of AI to enhance clear, compelling communication for different audiences."
My Response:
To address this, I included a prompt specifically recommending the use of Australian English for an Australian audience. I didn't embed this prompt directly in the system as it could compromise the quality of the content. Instead, users can provide this input as needed.
Competition Request:
"Ensure accuracy of content."
My Response:
I've taken multiple steps to ensure accuracy. The review process uses three LLAMA agents that independently assess the document. While accuracy can be enhanced by better models, the solution is model-agnostic, meaning we can easily switch to an improved model if needed.
Competition Request:
"Accessibility for inclusion."
My Response:
The Plain Language style guide discouraged using PDF output due to accessibility concerns. To address this, I chose to generate text files, which are accessible to a broader audience. I also incorporated multiple plain language style guides to ensure that the tool adheres to the widest possible set of plain language rules.
Competition Request:
"Support an agentic flow."
My Response:
This is a multi-agent system. One LLAMA agent processes the style guides and creates a ruleset, while three additional LLAMA agents apply these rules to the document to provide feedback.
Project Ideal Solutions and How I Addressed These:
Competition Request:
"Ideally, the solution should integrate with existing software like word processors, slides, and spreadsheets."
My Response:
The tool outputs in text format, allowing users to easily incorporate it into their workflows with word processors and other software.
Competition Request:
"Help users rewrite content in plain language."
My Response:
I thoroughly studied the Plain Language guidelines, its history, and key principles. This web app effectively helps users rewrite content to meet plain language standards.
Competition Request:
"Suggest inclusive language options."
My Response:
The LLAMA agents focus on different aspects of the document: style, word choice, and structure, which addresses inclusivity and language choice effectively.
Competition Request:
"Correct style, grammar, and punctuation errors."
My Response:
This is covered by the prompt provided to users. Embedding it within the system prompt could cause errors, so I left it as a separate user instruction.
Competition Request:
"Provide feedback to users so they can see document improvements and learn for next time."
My Response:
The web app provides highly detailed feedback, offering users insight into how their document was improved.