Latest from the team

Latest from the team

Latest from the team

Fresh Takes & Updates

Keep up with the latest feature rollouts, and updates from the team

All Posts

Team News

Changelog

Team News

May 18, 2025

Superpower Countdown: Ready, Set, Demo! - Part 3

As the highly anticipated engineers' conference rapidly approaches on the 24th of this month, our team is buzzing with excitement and dedicated efforts, steadily moving forward to unveil our groundbreaking AI-powered assistant, SuperPower.Engineer.

This blog post gives an in-depth look into our recent developments, showcasing the significant innovations, technical advancements, and enhancements we've achieved . . .

Bringing Ideas to Life: Interactive Dummy UI

To give life to our vision, we developed an engaging and interactive dummy  user interface using Streamlit. This rapid prototyping strategy enabled our team to experiment extensively, iteratively refining and optimising the interface.

Through continuous testing and feedback, we've created a seamless and intuitive user experience, closely aligning with real-world user expectations.

Our dummy / prototype UI has proven instrumental in identifying potential usability issues early, allowing us to enhance user interaction flows significantly before the final implementation.

A screenshot of a computer

AI-generated content may be incorrect.

Meme Magic: AI with a Sense of Humour

Injecting personality and creativity into our AI assistant, we introduced an engaging "Meme Mode," comprising 20 thoughtfully crafted, humorous prompts. This feature showcases Superpower’s playful side, adding an engaging dimension to user interactions.

The addition of Meme Mode underscores our commitment to providing a multifaceted AI experience that resonates with users, making interactions memorable, enjoyable, and relatable. This element of fun helps establish deeper user connections, fostering loyalty and satisfaction.

A person riding a unicorn

AI-generated content may be incorrect.

Deep Web Diving: Advanced Web Scraper

Understanding that a comprehensive knowledge base is fundamental to AI accuracy, our team developed a sophisticated web scraper capable of navigating through multiple layers of web pages.

This advanced scraper effectively captures detailed HTML content and extracts embedded PDFs, substantially enriching our assistant's knowledge repository.

The depth and breadth of data gathered enable Superpower to deliver precise, well-rounded, and highly informed responses, significantly improving its practical utility for users across diverse contexts.

What is Web Scraping? - Blog

Memory Mastery: DynamoDB Integration

Recognising the vital role of context in conversational AI, we implemented an advanced memory management system utilizing DynamoDB. This robust solution reliably tracks usernames, user IDs, conversation IDs, and chat session data, enhancing the continuity and personalization of user interactions.

By meticulously integrating user IDs and comprehensive session tracking, we've ensured that returning users experience seamless conversations, maintaining relevant context across multiple interactions.

This ensures a personalised, intuitive, and cohesive experience each time a user engages with our AI.

Seamless Session Integration: User ID and Chat Sessions

To further enhance personalisation and continuity, we've seamlessly integrated user ID and chat session tracking. This integration allows Superpower to recall previous interactions effortlessly, delivering context-aware, relevant, and personalised conversations.

Such an intuitive approach significantly improves user satisfaction, creating a consistently engaging and meaningful conversational experience. It also enables robust analytics, helping our team refine and continuously improve the AI’s performance and user interactions.

With the demo day fast approaching, excitement within our team continues to soar. Each accomplishment and innovation brings us one step closer to presenting an exceptional demonstration of Superpower's cutting-edge capabilities.

We’re confident our AI assistant will leave a lasting impression, showcasing the profound potential of technology-driven solutions. Stay tuned as we finalise our preparations—an extraordinary experience awaits!

Team News

May 18, 2025

Part 2: Giving our Rag Pipeline Superpowers: Memory, Management, and Ninja Mode

Integrating Memory into our LLM

Over the past two weeks, we've significantly upgraded our Retrieval-Augmented Generation (RAG) pipeline by integrating memory, ensuring our model remembers and utilizes previous interactions to enrich future responses.

By nature, Large Language Models (LLMs) are stateless. Each interaction stands alone without memory of previous conversations. To overcome this limitation, we've integrated a powerful framework offering multiple memory strategies:

  • Conversation Buffer Memory: Retains the entire conversation history, providing full context in subsequent interactions.

  • Conversation Buffer Window: Stores only a specific number of recent interactions, managing memory size efficiently.

  • Conversation Summary: Generates summaries of past conversations, allowing the model to retain context without overwhelming it with details.

For our specific use case, we opted for Conversation Buffer Memory, specifically retaining the last five conversation exchanges, balancing detailed memory with performance efficiency.

A computer on a table

AI-generated content may be incorrect.

Pinecone Integration and Multi-user Management

As we refined our RAG pipeline, we incorporated Pinecone, a test database, to systematically store conversation histories. Each interaction (user query and AI response) is indexed in Pinecone, effectively serving as a persistent memory store.

We advanced our setup further by assigning unique index within Pinecone for different users. Each user's conversation strings are independently stored within the user-allotted index under a specific namespace, which enhances user data management and security.

Tackling Cross-contamination in Document Retrieval

A critical issue we addressed was the potential cross-contamination, where the LLM could inadvertently mix documents from unrelated contexts to generate answers. Our solutions involved:

  1. User-Specific Concatenation: Identifying the user's specific system type (e.g. type of alarm system) and appending it directly to their queries. This ensured retrieval of only relevant documents tailored to the user's context.

  2. Namespace Segregation by Manufacturer: Creating separate namespaces within Pinecone for each alarm system manufacturer, further refining document retrieval and eliminating cross-contamination.

We're continuing extensive research to further optimize this aspect in the coming weeks.

Ninja Mode: Advanced Reasoning Activated

We introduced an exciting new feature, named Ninja Mode, transforming our AI assistant into an advanced troubleshooting expert. When activated, Ninja Mode leverages OpenAI’s state-of-the-art reasoning model, o1.

This advanced model enables the assistant to intelligently gather comprehensive contextual information, analyse complex situations methodically, and deliver highly precise troubleshooting insights.

Ninja Mode essentially provides our users with an expert-level AI technician capable of handling intricate queries with advanced reasoning capabilities.

Cartoon character in a black garment

AI-generated content may be incorrect.

Stay tuned as we continue to push the boundaries of AI capabilities, optimising our RAG pipeline further and exploring new frontiers in memory management and advanced AI troubleshooting!

Team News

May 18, 2025

Dev Team Progress update -  Part 1

Out of the gate we've primarily focused our resources on working out how to retrieve the most accurate answers possible from complex instruction manuals (our core product goal), by extracting & matching chunks of both text and image information - including the ability to extract an understanding of content embedded in images (which is - we found - a pretty complex challenge). 

Key progress to date includes:

  • Document Processing and RAG Pipeline Development:

    • Focused on refining document processing capabilities for a Retrieval-Augmented Generation (RAG) pipeline.

    • Conducted research and integrated a data structuring system to improve document parsing and processing - supporting data inputs from Amazon S3, with data storage in Pinecone vector database.

    • Enhanced data parsing by enabling:

      • Image extraction for detailed content descriptions.

      • Table extraction for structured data representation.

    • Implemented text chunking to improve similarity matching in the vector database.

    • Utilized OpenAI’s text-embedding-3-large model for vector embedding.

  • RAG Pipeline Creation:

    • Built and tested a RAG model and pipeline using Pinecone’s API and GPT-4o for answer generation.

    • Initial results demonstrated the pipeline's strong performance.

Following a review of the initial results we identified we were consistently generating accurate answers from in-depth complex manuals (which was the good news), however we needed to include wider contextual information to enrich and support the answers (e.g. including citations and more accurate / contextual image information).

Features & testing that followed included:

  • Citation Integration:

    • Incorporated source citations directly into LLM inputs to embed accurate citations within generated responses.

    • Enhanced the transparency, credibility, and traceability of the responses.

  • Detailed Source Referencing:

    • Extracted detailed metadata from Pinecone, including document IDs and specific page numbers.

    • Organized related documentation into an Amazon S3 bucket for easy access to original source PDFs.

    • Enabled users to verify facts by referencing the exact source pages.

  • Visual Guidance via Images:

    • Attempted to display images from referenced pages but faced limitations with metadata and base64 encoding.

    • Acknowledged the need for improved visual aids for complex questions (e.g., wiring diagrams) and planned for future refinement of this feature.

  • Incorporation of Internal Knowledge:

    • Integrated a structured dataset from a CSV file into the pipeline, including specific private knowledge sets like:

      • Terminology

      • Expert internal troubleshooting knowledge 

    • Improved the quality of generated responses by enriching them with highly accurate, contextually relevant internal knowledge.

  • Natural follow-up questions:

    • Added x3 follow-up questions per query, to dig deeper into the solution, for example, when programming is described at a top-level, the option to get the step-by-step programming information is provided.

Having made good progress retrieving accurate answers and providing citations (text and image) to help contextualise the answer, we're now moving onto dealing with more 'complex queries' - i.e. system-based queries that may involve more than a single 'instruction manual answer call', and involve various elements of more complex troubleshooting.

Hang tight for more progress and the next update - it's starting to get interesting ;)

Team News

May 18, 2025

Engineer field interviews: day 1

Abdur spent a day out on the road (in Easingwold, rural North Yorkshire) with Engineer Phil to help understand more first-hand the kinds of problems field-service engineers face out in the field. Observations from the day were:

  • YouTube: Engineer Phil's current No.1 trouble-shooting tool is YouTube. He said that he found the visual nature of it much more helpful than Google / instruction manuals, and referenced the ability to play / stop videos, to see how to fix an issue.

  • When trouble-shooting, Phil uses a mix of:

    • Experience / knowledge

    • Instruction manuals

    • Senior tech support (via phone)

  • Seemingly in most cases, if the instructions / problem-solve was available easily 'on tap' / 'on site' / 'on phone' Phil wouldn't need to speak to Tech Support (e.g. when Phil called colleague Asif to help with job 1, Asif's solution was to send an instructional diagram showing 'which wires / pins should be connected' to Phil - and Phil then followed this to solve his issue)

  • Phil highlighted that 'visual diagrams' / imagery is highly useful in the field (hence his love of YouTube)

  • For Abdur and Phil's 2nd customer visit, they encountered at 25 year-old vintage alarm system (the Paragon Super 2) This type alarm system is something that Phil had no specific experience of, so he immediately searched out the instruction manual on Google to find the answer. As a challenge, Abdur also used ChatGPT to find the answer - and ChatGPT proved a lot quicker in finding the answer, mostly as although Phil found the instruction manual - he then needed to trawl through multiple pages to try to find the right page - whereas ChatGPT served up an answer quickly on a plate.

Other insights from the day include:

  • Engineer Phil had never used ChatGPT before, and when we showed him it (i.e. fed a technical question to it, and then saw it spit out an answer) his jaw dropped. It was a magic moment.

  • Experience is seemingly as important as 'instructions from manuals' - as often things don't work as they should.

  • Methodical trouble-shooting is definitely a skill - but also definitely something that can be taught / explained (mechanically). Phil talked through a series of examples of how he has tackled problems in the past.

Team News

May 18, 2025

How do you know 2+2 = 4?

In the context of building a tool designed to share knowledge using Artificial Intelligence there's one fundamental question that's been bugging me, which is: 'how can a computer teach itself' - or to put it in a more grounded way: 'how does a computer know that 2+2=4 without being told / taught so by a third party (i.e. a human)?

Of course, the concept of knowledge and learning is no different when talking about machines as it is with humans, so the basic question is 'how do we learn stuff'? (and what is it to 'know' something?).

Listening to a podcast with Ilya Sutskever (one of the founders of OpenAI) he described very succinctly that humans learn in two ways:

  • Learning from others

  • Learning by trial and error (& patterns)

So in the context of a computer not being able to 'learn from someone else' then the only other way it can learn is by trial and error - or a more smarter version of trial and error being 'pattern matching'. For example, the computer may observe that 1+1 = 2, and therefore guess that the pattern follows that 2+2=4.

It's this 'guessing' that underpins the intelligence and knowledge of AI. Essentially all AI is doing is a very sophisticated very of guessing (aka probability) - and thanks to extremely large neural networks powering Large Language Models (LLMs) this 'guessing' is next-level guessing (aka highly technical and increasingly accurate probability modelling).

Other than listening to AI OG's like Ilya Sutskever for insights into how humans and machines learn, I've also started quizzing my 8 year-old daughter for insights into learning - as she is a teaching obsessive and has a healthy love of writing school lesson plans in her spare time.

Interestingly, her answer to the 'how do we know 2+2 = 4' question was bang on Sutskever's answer - telling me: 'Well Daddy, it's basic maths if 1+1 = 2 then of course 2+2 = 4'.

The simple pattern matching of an 8 year old's brain underlines a fundamental truth of AI - that all intelligence is essentially 'guessing' (or a very accurate version of guessing based on probabilistic pattern-matching).

Team News

May 18, 2025

Aiming for 100% Artificial - starting with our Logo

As a core goal of SuperPower.Engineer is to prove the transformative power of AI on the physical work environment, we also thought it would be fun (and useful / efficient) to shape the wider project using as much AI tech as possible - starting with the Logo.

As of today, AI is 80 to 99% good at a lot of tasks. Logo (image) generation I'd say is around the 85% mark.

Here's the journey we took in creating the SuperPower logo - using a mix of ChatGPT and human design talent.

Step 1: ChatGPT prompt 1

The first thing we did was feed a prompt into ChatGPT along the lines 'Make us a logo for our new AI start-up called SuperPower. . . '. (You can read the full prompt & process via ChatGPT here: https://chatgpt.com/share/67b5f6f1-a0e4-800b-85f3-2d7a6ee8247d)

Within about 60 seconds - after a bit of questioning re: some design specifics - ChatGPT delivered an initial design, which looked like this (which I'd say was about 20% good):

Step 2: ChatGPT prompt 2

Mildly unhappy with the output, I then gave ChatGPT a further prompt to include 'SP' in the middle of the logo (which was unclear in the original), along with the word 'SuperPower', and miraculously it spat out the following - which I'd say was about 85% good:

Step 3: Dave the designer

To get it to 100% good, I decided to ditch ChatGPT and pass over the v1 design to my erstwhile designer friend Dave - who then took another c.30 mins to tweak it & turn it into what we have now.

Here's the final stage brief I have to Dave - using the core initial design from ChatGPT:

And here's the final design output then cleaned-up and created by human designer Dave:

The whole process (from a my time perspective) realistically took about 15 mins (which was basically my briefing up Dave with the basics that ChatGPT created), and another 30 mins from Dave - meaning end-to-end, our logo took about 45 human minutes to create. With a rough 60% AI and 40% human input ratio.

All Posts

Team News

Changelog

Team News

May 18, 2025

Superpower Countdown: Ready, Set, Demo! - Part 3

As the highly anticipated engineers' conference rapidly approaches on the 24th of this month, our team is buzzing with excitement and dedicated efforts, steadily moving forward to unveil our groundbreaking AI-powered assistant, SuperPower.Engineer.

This blog post gives an in-depth look into our recent developments, showcasing the significant innovations, technical advancements, and enhancements we've achieved . . .

Bringing Ideas to Life: Interactive Dummy UI

To give life to our vision, we developed an engaging and interactive dummy  user interface using Streamlit. This rapid prototyping strategy enabled our team to experiment extensively, iteratively refining and optimising the interface.

Through continuous testing and feedback, we've created a seamless and intuitive user experience, closely aligning with real-world user expectations.

Our dummy / prototype UI has proven instrumental in identifying potential usability issues early, allowing us to enhance user interaction flows significantly before the final implementation.

A screenshot of a computer

AI-generated content may be incorrect.

Meme Magic: AI with a Sense of Humour

Injecting personality and creativity into our AI assistant, we introduced an engaging "Meme Mode," comprising 20 thoughtfully crafted, humorous prompts. This feature showcases Superpower’s playful side, adding an engaging dimension to user interactions.

The addition of Meme Mode underscores our commitment to providing a multifaceted AI experience that resonates with users, making interactions memorable, enjoyable, and relatable. This element of fun helps establish deeper user connections, fostering loyalty and satisfaction.

A person riding a unicorn

AI-generated content may be incorrect.

Deep Web Diving: Advanced Web Scraper

Understanding that a comprehensive knowledge base is fundamental to AI accuracy, our team developed a sophisticated web scraper capable of navigating through multiple layers of web pages.

This advanced scraper effectively captures detailed HTML content and extracts embedded PDFs, substantially enriching our assistant's knowledge repository.

The depth and breadth of data gathered enable Superpower to deliver precise, well-rounded, and highly informed responses, significantly improving its practical utility for users across diverse contexts.

What is Web Scraping? - Blog

Memory Mastery: DynamoDB Integration

Recognising the vital role of context in conversational AI, we implemented an advanced memory management system utilizing DynamoDB. This robust solution reliably tracks usernames, user IDs, conversation IDs, and chat session data, enhancing the continuity and personalization of user interactions.

By meticulously integrating user IDs and comprehensive session tracking, we've ensured that returning users experience seamless conversations, maintaining relevant context across multiple interactions.

This ensures a personalised, intuitive, and cohesive experience each time a user engages with our AI.

Seamless Session Integration: User ID and Chat Sessions

To further enhance personalisation and continuity, we've seamlessly integrated user ID and chat session tracking. This integration allows Superpower to recall previous interactions effortlessly, delivering context-aware, relevant, and personalised conversations.

Such an intuitive approach significantly improves user satisfaction, creating a consistently engaging and meaningful conversational experience. It also enables robust analytics, helping our team refine and continuously improve the AI’s performance and user interactions.

With the demo day fast approaching, excitement within our team continues to soar. Each accomplishment and innovation brings us one step closer to presenting an exceptional demonstration of Superpower's cutting-edge capabilities.

We’re confident our AI assistant will leave a lasting impression, showcasing the profound potential of technology-driven solutions. Stay tuned as we finalise our preparations—an extraordinary experience awaits!

Team News

May 18, 2025

Part 2: Giving our Rag Pipeline Superpowers: Memory, Management, and Ninja Mode

Integrating Memory into our LLM

Over the past two weeks, we've significantly upgraded our Retrieval-Augmented Generation (RAG) pipeline by integrating memory, ensuring our model remembers and utilizes previous interactions to enrich future responses.

By nature, Large Language Models (LLMs) are stateless. Each interaction stands alone without memory of previous conversations. To overcome this limitation, we've integrated a powerful framework offering multiple memory strategies:

  • Conversation Buffer Memory: Retains the entire conversation history, providing full context in subsequent interactions.

  • Conversation Buffer Window: Stores only a specific number of recent interactions, managing memory size efficiently.

  • Conversation Summary: Generates summaries of past conversations, allowing the model to retain context without overwhelming it with details.

For our specific use case, we opted for Conversation Buffer Memory, specifically retaining the last five conversation exchanges, balancing detailed memory with performance efficiency.

A computer on a table

AI-generated content may be incorrect.

Pinecone Integration and Multi-user Management

As we refined our RAG pipeline, we incorporated Pinecone, a test database, to systematically store conversation histories. Each interaction (user query and AI response) is indexed in Pinecone, effectively serving as a persistent memory store.

We advanced our setup further by assigning unique index within Pinecone for different users. Each user's conversation strings are independently stored within the user-allotted index under a specific namespace, which enhances user data management and security.

Tackling Cross-contamination in Document Retrieval

A critical issue we addressed was the potential cross-contamination, where the LLM could inadvertently mix documents from unrelated contexts to generate answers. Our solutions involved:

  1. User-Specific Concatenation: Identifying the user's specific system type (e.g. type of alarm system) and appending it directly to their queries. This ensured retrieval of only relevant documents tailored to the user's context.

  2. Namespace Segregation by Manufacturer: Creating separate namespaces within Pinecone for each alarm system manufacturer, further refining document retrieval and eliminating cross-contamination.

We're continuing extensive research to further optimize this aspect in the coming weeks.

Ninja Mode: Advanced Reasoning Activated

We introduced an exciting new feature, named Ninja Mode, transforming our AI assistant into an advanced troubleshooting expert. When activated, Ninja Mode leverages OpenAI’s state-of-the-art reasoning model, o1.

This advanced model enables the assistant to intelligently gather comprehensive contextual information, analyse complex situations methodically, and deliver highly precise troubleshooting insights.

Ninja Mode essentially provides our users with an expert-level AI technician capable of handling intricate queries with advanced reasoning capabilities.

Cartoon character in a black garment

AI-generated content may be incorrect.

Stay tuned as we continue to push the boundaries of AI capabilities, optimising our RAG pipeline further and exploring new frontiers in memory management and advanced AI troubleshooting!

Team News

May 18, 2025

Dev Team Progress update -  Part 1

Out of the gate we've primarily focused our resources on working out how to retrieve the most accurate answers possible from complex instruction manuals (our core product goal), by extracting & matching chunks of both text and image information - including the ability to extract an understanding of content embedded in images (which is - we found - a pretty complex challenge). 

Key progress to date includes:

  • Document Processing and RAG Pipeline Development:

    • Focused on refining document processing capabilities for a Retrieval-Augmented Generation (RAG) pipeline.

    • Conducted research and integrated a data structuring system to improve document parsing and processing - supporting data inputs from Amazon S3, with data storage in Pinecone vector database.

    • Enhanced data parsing by enabling:

      • Image extraction for detailed content descriptions.

      • Table extraction for structured data representation.

    • Implemented text chunking to improve similarity matching in the vector database.

    • Utilized OpenAI’s text-embedding-3-large model for vector embedding.

  • RAG Pipeline Creation:

    • Built and tested a RAG model and pipeline using Pinecone’s API and GPT-4o for answer generation.

    • Initial results demonstrated the pipeline's strong performance.

Following a review of the initial results we identified we were consistently generating accurate answers from in-depth complex manuals (which was the good news), however we needed to include wider contextual information to enrich and support the answers (e.g. including citations and more accurate / contextual image information).

Features & testing that followed included:

  • Citation Integration:

    • Incorporated source citations directly into LLM inputs to embed accurate citations within generated responses.

    • Enhanced the transparency, credibility, and traceability of the responses.

  • Detailed Source Referencing:

    • Extracted detailed metadata from Pinecone, including document IDs and specific page numbers.

    • Organized related documentation into an Amazon S3 bucket for easy access to original source PDFs.

    • Enabled users to verify facts by referencing the exact source pages.

  • Visual Guidance via Images:

    • Attempted to display images from referenced pages but faced limitations with metadata and base64 encoding.

    • Acknowledged the need for improved visual aids for complex questions (e.g., wiring diagrams) and planned for future refinement of this feature.

  • Incorporation of Internal Knowledge:

    • Integrated a structured dataset from a CSV file into the pipeline, including specific private knowledge sets like:

      • Terminology

      • Expert internal troubleshooting knowledge 

    • Improved the quality of generated responses by enriching them with highly accurate, contextually relevant internal knowledge.

  • Natural follow-up questions:

    • Added x3 follow-up questions per query, to dig deeper into the solution, for example, when programming is described at a top-level, the option to get the step-by-step programming information is provided.

Having made good progress retrieving accurate answers and providing citations (text and image) to help contextualise the answer, we're now moving onto dealing with more 'complex queries' - i.e. system-based queries that may involve more than a single 'instruction manual answer call', and involve various elements of more complex troubleshooting.

Hang tight for more progress and the next update - it's starting to get interesting ;)

Team News

May 18, 2025

Engineer field interviews: day 1

Abdur spent a day out on the road (in Easingwold, rural North Yorkshire) with Engineer Phil to help understand more first-hand the kinds of problems field-service engineers face out in the field. Observations from the day were:

  • YouTube: Engineer Phil's current No.1 trouble-shooting tool is YouTube. He said that he found the visual nature of it much more helpful than Google / instruction manuals, and referenced the ability to play / stop videos, to see how to fix an issue.

  • When trouble-shooting, Phil uses a mix of:

    • Experience / knowledge

    • Instruction manuals

    • Senior tech support (via phone)

  • Seemingly in most cases, if the instructions / problem-solve was available easily 'on tap' / 'on site' / 'on phone' Phil wouldn't need to speak to Tech Support (e.g. when Phil called colleague Asif to help with job 1, Asif's solution was to send an instructional diagram showing 'which wires / pins should be connected' to Phil - and Phil then followed this to solve his issue)

  • Phil highlighted that 'visual diagrams' / imagery is highly useful in the field (hence his love of YouTube)

  • For Abdur and Phil's 2nd customer visit, they encountered at 25 year-old vintage alarm system (the Paragon Super 2) This type alarm system is something that Phil had no specific experience of, so he immediately searched out the instruction manual on Google to find the answer. As a challenge, Abdur also used ChatGPT to find the answer - and ChatGPT proved a lot quicker in finding the answer, mostly as although Phil found the instruction manual - he then needed to trawl through multiple pages to try to find the right page - whereas ChatGPT served up an answer quickly on a plate.

Other insights from the day include:

  • Engineer Phil had never used ChatGPT before, and when we showed him it (i.e. fed a technical question to it, and then saw it spit out an answer) his jaw dropped. It was a magic moment.

  • Experience is seemingly as important as 'instructions from manuals' - as often things don't work as they should.

  • Methodical trouble-shooting is definitely a skill - but also definitely something that can be taught / explained (mechanically). Phil talked through a series of examples of how he has tackled problems in the past.

Team News

May 18, 2025

How do you know 2+2 = 4?

In the context of building a tool designed to share knowledge using Artificial Intelligence there's one fundamental question that's been bugging me, which is: 'how can a computer teach itself' - or to put it in a more grounded way: 'how does a computer know that 2+2=4 without being told / taught so by a third party (i.e. a human)?

Of course, the concept of knowledge and learning is no different when talking about machines as it is with humans, so the basic question is 'how do we learn stuff'? (and what is it to 'know' something?).

Listening to a podcast with Ilya Sutskever (one of the founders of OpenAI) he described very succinctly that humans learn in two ways:

  • Learning from others

  • Learning by trial and error (& patterns)

So in the context of a computer not being able to 'learn from someone else' then the only other way it can learn is by trial and error - or a more smarter version of trial and error being 'pattern matching'. For example, the computer may observe that 1+1 = 2, and therefore guess that the pattern follows that 2+2=4.

It's this 'guessing' that underpins the intelligence and knowledge of AI. Essentially all AI is doing is a very sophisticated very of guessing (aka probability) - and thanks to extremely large neural networks powering Large Language Models (LLMs) this 'guessing' is next-level guessing (aka highly technical and increasingly accurate probability modelling).

Other than listening to AI OG's like Ilya Sutskever for insights into how humans and machines learn, I've also started quizzing my 8 year-old daughter for insights into learning - as she is a teaching obsessive and has a healthy love of writing school lesson plans in her spare time.

Interestingly, her answer to the 'how do we know 2+2 = 4' question was bang on Sutskever's answer - telling me: 'Well Daddy, it's basic maths if 1+1 = 2 then of course 2+2 = 4'.

The simple pattern matching of an 8 year old's brain underlines a fundamental truth of AI - that all intelligence is essentially 'guessing' (or a very accurate version of guessing based on probabilistic pattern-matching).

Team News

May 18, 2025

Aiming for 100% Artificial - starting with our Logo

As a core goal of SuperPower.Engineer is to prove the transformative power of AI on the physical work environment, we also thought it would be fun (and useful / efficient) to shape the wider project using as much AI tech as possible - starting with the Logo.

As of today, AI is 80 to 99% good at a lot of tasks. Logo (image) generation I'd say is around the 85% mark.

Here's the journey we took in creating the SuperPower logo - using a mix of ChatGPT and human design talent.

Step 1: ChatGPT prompt 1

The first thing we did was feed a prompt into ChatGPT along the lines 'Make us a logo for our new AI start-up called SuperPower. . . '. (You can read the full prompt & process via ChatGPT here: https://chatgpt.com/share/67b5f6f1-a0e4-800b-85f3-2d7a6ee8247d)

Within about 60 seconds - after a bit of questioning re: some design specifics - ChatGPT delivered an initial design, which looked like this (which I'd say was about 20% good):

Step 2: ChatGPT prompt 2

Mildly unhappy with the output, I then gave ChatGPT a further prompt to include 'SP' in the middle of the logo (which was unclear in the original), along with the word 'SuperPower', and miraculously it spat out the following - which I'd say was about 85% good:

Step 3: Dave the designer

To get it to 100% good, I decided to ditch ChatGPT and pass over the v1 design to my erstwhile designer friend Dave - who then took another c.30 mins to tweak it & turn it into what we have now.

Here's the final stage brief I have to Dave - using the core initial design from ChatGPT:

And here's the final design output then cleaned-up and created by human designer Dave:

The whole process (from a my time perspective) realistically took about 15 mins (which was basically my briefing up Dave with the basics that ChatGPT created), and another 30 mins from Dave - meaning end-to-end, our logo took about 45 human minutes to create. With a rough 60% AI and 40% human input ratio.

All Posts

Thinking

The rise of the Blue Collar worker in the age of AI

Chatting to Michael (a friend-of-a-friend)  in my local South London pub, we discussed the rapid changing landscape of Soho.  Michael is a recently out-of-work film-maker, struggling because of the rise of AI.  His skills are no longer needed, replaced by Midjourney, Leonardo, Google's Flow  (and the long list of AI-powered tools). 

Michael is not alone it seems, as news broke that the once-famous behemoth Special FX house  The Mill was shutting shop in Soho.  Its need similarly evaporated by AI prompts.

20 year Soho flashback

This loss of jobs in Soho reminded me of a similar vast loss of jobs & change that happened in Soho 20 years back.  The loss back then was of the Bicycle Courier.

Back in the late 90's and early 2000's when I first started working in the area, Soho was buzzing with armies of fixie-riding couriers, delivering manuscripts, reels, tapes and hard drives from office to office. 

The courier formed a core part of the Soho information super highway - delivering (by hand and on wheels) information from one office to another.  The courier was also a core part of Soho culture, propping up bars in pubs (the Coach & Horses on Greek Street was a favourite) along with strip clubs and brothels that were ubiquitous at the time.

Broadband internet drove this mass cull of jobs.  At the click of a button you could send a film manuscript, in the same way 20 years later a click can now not only send a manuscript but also write the manuscript in the first place.

AI-proof 'Trades'

And just like the army of couriers 20 years ago, Michael and his cohort are forced to make a career change with many now considering life in the 'Trades'.

Film-makers like couriers are generally quite practical people, so a move into becoming an electrician or plumber is a logical move for many. 

Importantly, Blue Collar manual jobs can't easily be decimated by AI, and in fact are in a position to thrive in an AI era, thanks to tools like Super Engineer.

At Super Engineer we see ourselves as champions of the Trades, using AI to help protect & scale jobs in the Blue Collar space, helping up-skill & empower workers using the power of AI - especially in the space of Technical Services & service engineers (like electricians, plumbers, HVAC, fire & security, Facility Managers).

Helping make Service Engineers Smarter using AI

One of the biggest challenges faced by all Technical Service Engineers isn't the threat of AI on their jobs, but the rising complexity of technologies and the breadth of technologies engineers are expected to deal with & fix.

Put simply, technology has got infinitely more complex but engineers haven't got infinitely smarter.

That's where AI can help.

The future = AI glasses (and more!)

Imagine a world where the engineer visiting your house to fix your boiler, doesn't have to spend 30 minutes on the phone to  the Equipment Manufacturer's Tech Support line to understand what a certain Error Code is & how to fix it - as it's a piece of equipment he's unfamiliar with.

Instead, the engineer pops on his AR glasses.  The camera on the glasses identifies what the specific piece of equipment is and with the help of AI then identifies what the error code means, locates the specific page in the equipment manual to give full context and then suggests a step-by-step way to fix the problem - with instructions being to the engineer via an earpiece & also via a head-up-display on the glasses (if required).

Problem fixed.  Customer happy.  Engineer happy.

Welcome to the future of AI-empowered Service Engineers, a world where AI adds huge value & skills to workers rather than replacing them.

Welcome to our mission at Super Engineer ;-)

Team News

May 18, 2025

Superpower Countdown: Ready, Set, Demo! - Part 3

As the highly anticipated engineers' conference rapidly approaches on the 24th of this month, our team is buzzing with excitement and dedicated efforts, steadily moving forward to unveil our groundbreaking AI-powered assistant, SuperPower.Engineer.

This blog post gives an in-depth look into our recent developments, showcasing the significant innovations, technical advancements, and enhancements we've achieved . . .

Bringing Ideas to Life: Interactive Dummy UI

To give life to our vision, we developed an engaging and interactive dummy  user interface using Streamlit. This rapid prototyping strategy enabled our team to experiment extensively, iteratively refining and optimising the interface.

Through continuous testing and feedback, we've created a seamless and intuitive user experience, closely aligning with real-world user expectations.

Our dummy / prototype UI has proven instrumental in identifying potential usability issues early, allowing us to enhance user interaction flows significantly before the final implementation.

A screenshot of a computer

AI-generated content may be incorrect.

Meme Magic: AI with a Sense of Humour

Injecting personality and creativity into our AI assistant, we introduced an engaging "Meme Mode," comprising 20 thoughtfully crafted, humorous prompts. This feature showcases Superpower’s playful side, adding an engaging dimension to user interactions.

The addition of Meme Mode underscores our commitment to providing a multifaceted AI experience that resonates with users, making interactions memorable, enjoyable, and relatable. This element of fun helps establish deeper user connections, fostering loyalty and satisfaction.

A person riding a unicorn

AI-generated content may be incorrect.

Deep Web Diving: Advanced Web Scraper

Understanding that a comprehensive knowledge base is fundamental to AI accuracy, our team developed a sophisticated web scraper capable of navigating through multiple layers of web pages.

This advanced scraper effectively captures detailed HTML content and extracts embedded PDFs, substantially enriching our assistant's knowledge repository.

The depth and breadth of data gathered enable Superpower to deliver precise, well-rounded, and highly informed responses, significantly improving its practical utility for users across diverse contexts.

What is Web Scraping? - Blog

Memory Mastery: DynamoDB Integration

Recognising the vital role of context in conversational AI, we implemented an advanced memory management system utilizing DynamoDB. This robust solution reliably tracks usernames, user IDs, conversation IDs, and chat session data, enhancing the continuity and personalization of user interactions.

By meticulously integrating user IDs and comprehensive session tracking, we've ensured that returning users experience seamless conversations, maintaining relevant context across multiple interactions.

This ensures a personalised, intuitive, and cohesive experience each time a user engages with our AI.

Seamless Session Integration: User ID and Chat Sessions

To further enhance personalisation and continuity, we've seamlessly integrated user ID and chat session tracking. This integration allows Superpower to recall previous interactions effortlessly, delivering context-aware, relevant, and personalised conversations.

Such an intuitive approach significantly improves user satisfaction, creating a consistently engaging and meaningful conversational experience. It also enables robust analytics, helping our team refine and continuously improve the AI’s performance and user interactions.

With the demo day fast approaching, excitement within our team continues to soar. Each accomplishment and innovation brings us one step closer to presenting an exceptional demonstration of Superpower's cutting-edge capabilities.

We’re confident our AI assistant will leave a lasting impression, showcasing the profound potential of technology-driven solutions. Stay tuned as we finalise our preparations—an extraordinary experience awaits!

Team News

May 18, 2025

Part 2: Giving our Rag Pipeline Superpowers: Memory, Management, and Ninja Mode

Integrating Memory into our LLM

Over the past two weeks, we've significantly upgraded our Retrieval-Augmented Generation (RAG) pipeline by integrating memory, ensuring our model remembers and utilizes previous interactions to enrich future responses.

By nature, Large Language Models (LLMs) are stateless. Each interaction stands alone without memory of previous conversations. To overcome this limitation, we've integrated a powerful framework offering multiple memory strategies:

  • Conversation Buffer Memory: Retains the entire conversation history, providing full context in subsequent interactions.

  • Conversation Buffer Window: Stores only a specific number of recent interactions, managing memory size efficiently.

  • Conversation Summary: Generates summaries of past conversations, allowing the model to retain context without overwhelming it with details.

For our specific use case, we opted for Conversation Buffer Memory, specifically retaining the last five conversation exchanges, balancing detailed memory with performance efficiency.

A computer on a table

AI-generated content may be incorrect.

Pinecone Integration and Multi-user Management

As we refined our RAG pipeline, we incorporated Pinecone, a test database, to systematically store conversation histories. Each interaction (user query and AI response) is indexed in Pinecone, effectively serving as a persistent memory store.

We advanced our setup further by assigning unique index within Pinecone for different users. Each user's conversation strings are independently stored within the user-allotted index under a specific namespace, which enhances user data management and security.

Tackling Cross-contamination in Document Retrieval

A critical issue we addressed was the potential cross-contamination, where the LLM could inadvertently mix documents from unrelated contexts to generate answers. Our solutions involved:

  1. User-Specific Concatenation: Identifying the user's specific system type (e.g. type of alarm system) and appending it directly to their queries. This ensured retrieval of only relevant documents tailored to the user's context.

  2. Namespace Segregation by Manufacturer: Creating separate namespaces within Pinecone for each alarm system manufacturer, further refining document retrieval and eliminating cross-contamination.

We're continuing extensive research to further optimize this aspect in the coming weeks.

Ninja Mode: Advanced Reasoning Activated

We introduced an exciting new feature, named Ninja Mode, transforming our AI assistant into an advanced troubleshooting expert. When activated, Ninja Mode leverages OpenAI’s state-of-the-art reasoning model, o1.

This advanced model enables the assistant to intelligently gather comprehensive contextual information, analyse complex situations methodically, and deliver highly precise troubleshooting insights.

Ninja Mode essentially provides our users with an expert-level AI technician capable of handling intricate queries with advanced reasoning capabilities.

Cartoon character in a black garment

AI-generated content may be incorrect.

Stay tuned as we continue to push the boundaries of AI capabilities, optimising our RAG pipeline further and exploring new frontiers in memory management and advanced AI troubleshooting!

Team News

May 18, 2025

Dev Team Progress update -  Part 1

Out of the gate we've primarily focused our resources on working out how to retrieve the most accurate answers possible from complex instruction manuals (our core product goal), by extracting & matching chunks of both text and image information - including the ability to extract an understanding of content embedded in images (which is - we found - a pretty complex challenge). 

Key progress to date includes:

  • Document Processing and RAG Pipeline Development:

    • Focused on refining document processing capabilities for a Retrieval-Augmented Generation (RAG) pipeline.

    • Conducted research and integrated a data structuring system to improve document parsing and processing - supporting data inputs from Amazon S3, with data storage in Pinecone vector database.

    • Enhanced data parsing by enabling:

      • Image extraction for detailed content descriptions.

      • Table extraction for structured data representation.

    • Implemented text chunking to improve similarity matching in the vector database.

    • Utilized OpenAI’s text-embedding-3-large model for vector embedding.

  • RAG Pipeline Creation:

    • Built and tested a RAG model and pipeline using Pinecone’s API and GPT-4o for answer generation.

    • Initial results demonstrated the pipeline's strong performance.

Following a review of the initial results we identified we were consistently generating accurate answers from in-depth complex manuals (which was the good news), however we needed to include wider contextual information to enrich and support the answers (e.g. including citations and more accurate / contextual image information).

Features & testing that followed included:

  • Citation Integration:

    • Incorporated source citations directly into LLM inputs to embed accurate citations within generated responses.

    • Enhanced the transparency, credibility, and traceability of the responses.

  • Detailed Source Referencing:

    • Extracted detailed metadata from Pinecone, including document IDs and specific page numbers.

    • Organized related documentation into an Amazon S3 bucket for easy access to original source PDFs.

    • Enabled users to verify facts by referencing the exact source pages.

  • Visual Guidance via Images:

    • Attempted to display images from referenced pages but faced limitations with metadata and base64 encoding.

    • Acknowledged the need for improved visual aids for complex questions (e.g., wiring diagrams) and planned for future refinement of this feature.

  • Incorporation of Internal Knowledge:

    • Integrated a structured dataset from a CSV file into the pipeline, including specific private knowledge sets like:

      • Terminology

      • Expert internal troubleshooting knowledge 

    • Improved the quality of generated responses by enriching them with highly accurate, contextually relevant internal knowledge.

  • Natural follow-up questions:

    • Added x3 follow-up questions per query, to dig deeper into the solution, for example, when programming is described at a top-level, the option to get the step-by-step programming information is provided.

Having made good progress retrieving accurate answers and providing citations (text and image) to help contextualise the answer, we're now moving onto dealing with more 'complex queries' - i.e. system-based queries that may involve more than a single 'instruction manual answer call', and involve various elements of more complex troubleshooting.

Hang tight for more progress and the next update - it's starting to get interesting ;)

Team News

May 18, 2025

Engineer field interviews: day 1

Abdur spent a day out on the road (in Easingwold, rural North Yorkshire) with Engineer Phil to help understand more first-hand the kinds of problems field-service engineers face out in the field. Observations from the day were:

  • YouTube: Engineer Phil's current No.1 trouble-shooting tool is YouTube. He said that he found the visual nature of it much more helpful than Google / instruction manuals, and referenced the ability to play / stop videos, to see how to fix an issue.

  • When trouble-shooting, Phil uses a mix of:

    • Experience / knowledge

    • Instruction manuals

    • Senior tech support (via phone)

  • Seemingly in most cases, if the instructions / problem-solve was available easily 'on tap' / 'on site' / 'on phone' Phil wouldn't need to speak to Tech Support (e.g. when Phil called colleague Asif to help with job 1, Asif's solution was to send an instructional diagram showing 'which wires / pins should be connected' to Phil - and Phil then followed this to solve his issue)

  • Phil highlighted that 'visual diagrams' / imagery is highly useful in the field (hence his love of YouTube)

  • For Abdur and Phil's 2nd customer visit, they encountered at 25 year-old vintage alarm system (the Paragon Super 2) This type alarm system is something that Phil had no specific experience of, so he immediately searched out the instruction manual on Google to find the answer. As a challenge, Abdur also used ChatGPT to find the answer - and ChatGPT proved a lot quicker in finding the answer, mostly as although Phil found the instruction manual - he then needed to trawl through multiple pages to try to find the right page - whereas ChatGPT served up an answer quickly on a plate.

Other insights from the day include:

  • Engineer Phil had never used ChatGPT before, and when we showed him it (i.e. fed a technical question to it, and then saw it spit out an answer) his jaw dropped. It was a magic moment.

  • Experience is seemingly as important as 'instructions from manuals' - as often things don't work as they should.

  • Methodical trouble-shooting is definitely a skill - but also definitely something that can be taught / explained (mechanically). Phil talked through a series of examples of how he has tackled problems in the past.

Team News

May 18, 2025

How do you know 2+2 = 4?

In the context of building a tool designed to share knowledge using Artificial Intelligence there's one fundamental question that's been bugging me, which is: 'how can a computer teach itself' - or to put it in a more grounded way: 'how does a computer know that 2+2=4 without being told / taught so by a third party (i.e. a human)?

Of course, the concept of knowledge and learning is no different when talking about machines as it is with humans, so the basic question is 'how do we learn stuff'? (and what is it to 'know' something?).

Listening to a podcast with Ilya Sutskever (one of the founders of OpenAI) he described very succinctly that humans learn in two ways:

  • Learning from others

  • Learning by trial and error (& patterns)

So in the context of a computer not being able to 'learn from someone else' then the only other way it can learn is by trial and error - or a more smarter version of trial and error being 'pattern matching'. For example, the computer may observe that 1+1 = 2, and therefore guess that the pattern follows that 2+2=4.

It's this 'guessing' that underpins the intelligence and knowledge of AI. Essentially all AI is doing is a very sophisticated very of guessing (aka probability) - and thanks to extremely large neural networks powering Large Language Models (LLMs) this 'guessing' is next-level guessing (aka highly technical and increasingly accurate probability modelling).

Other than listening to AI OG's like Ilya Sutskever for insights into how humans and machines learn, I've also started quizzing my 8 year-old daughter for insights into learning - as she is a teaching obsessive and has a healthy love of writing school lesson plans in her spare time.

Interestingly, her answer to the 'how do we know 2+2 = 4' question was bang on Sutskever's answer - telling me: 'Well Daddy, it's basic maths if 1+1 = 2 then of course 2+2 = 4'.

The simple pattern matching of an 8 year old's brain underlines a fundamental truth of AI - that all intelligence is essentially 'guessing' (or a very accurate version of guessing based on probabilistic pattern-matching).

Team News

May 18, 2025

Aiming for 100% Artificial - starting with our Logo

As a core goal of SuperPower.Engineer is to prove the transformative power of AI on the physical work environment, we also thought it would be fun (and useful / efficient) to shape the wider project using as much AI tech as possible - starting with the Logo.

As of today, AI is 80 to 99% good at a lot of tasks. Logo (image) generation I'd say is around the 85% mark.

Here's the journey we took in creating the SuperPower logo - using a mix of ChatGPT and human design talent.

Step 1: ChatGPT prompt 1

The first thing we did was feed a prompt into ChatGPT along the lines 'Make us a logo for our new AI start-up called SuperPower. . . '. (You can read the full prompt & process via ChatGPT here: https://chatgpt.com/share/67b5f6f1-a0e4-800b-85f3-2d7a6ee8247d)

Within about 60 seconds - after a bit of questioning re: some design specifics - ChatGPT delivered an initial design, which looked like this (which I'd say was about 20% good):

Step 2: ChatGPT prompt 2

Mildly unhappy with the output, I then gave ChatGPT a further prompt to include 'SP' in the middle of the logo (which was unclear in the original), along with the word 'SuperPower', and miraculously it spat out the following - which I'd say was about 85% good:

Step 3: Dave the designer

To get it to 100% good, I decided to ditch ChatGPT and pass over the v1 design to my erstwhile designer friend Dave - who then took another c.30 mins to tweak it & turn it into what we have now.

Here's the final stage brief I have to Dave - using the core initial design from ChatGPT:

And here's the final design output then cleaned-up and created by human designer Dave:

The whole process (from a my time perspective) realistically took about 15 mins (which was basically my briefing up Dave with the basics that ChatGPT created), and another 30 mins from Dave - meaning end-to-end, our logo took about 45 human minutes to create. With a rough 60% AI and 40% human input ratio.

Load More