top of page
LATEST TECH ARTICLES
![[On-Device AI Chatbot] Part 10: The Future of On-Device AI and TecAce's Roadmap (Conclusion)](https://static.wixstatic.com/media/2ea07e_d1771a9889764093a8c855756693ba51~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_d1771a9889764093a8c855756693ba51~mv2.webp)
![[On-Device AI Chatbot] Part 10: The Future of On-Device AI and TecAce's Roadmap (Conclusion)](https://static.wixstatic.com/media/2ea07e_d1771a9889764093a8c855756693ba51~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_d1771a9889764093a8c855756693ba51~mv2.webp)
[On-Device AI Chatbot] Part 10: The Future of On-Device AI and TecAce's Roadmap (Conclusion)
The Future of On-Device AI and TecAce's Roadmap Throughout this 9-part series, we have chronicled the entire journey of developing an on-device chatbot—a solution to cloud cost and data security issues. We covered everything from selecting a Small Language Model (SLM) and applying quantization, integrating offline STT/TTS, building local RAG, to rigorously validating quality using AI SuperVision and overcoming hardware performance constraints. In this grand finale, Part 10,
7 days ago
![[On-Device AI Chatbot] Part 9: Challenging Performance Limits: Heat, Battery, and Response Speed](https://static.wixstatic.com/media/2ea07e_826bc45db874477090ea018335b34059~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_826bc45db874477090ea018335b34059~mv2.webp)
![[On-Device AI Chatbot] Part 9: Challenging Performance Limits: Heat, Battery, and Response Speed](https://static.wixstatic.com/media/2ea07e_826bc45db874477090ea018335b34059~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_826bc45db874477090ea018335b34059~mv2.webp)
[On-Device AI Chatbot] Part 9: Challenging Performance Limits: Heat, Battery, and Response Speed
Challenging Performance Limits Heat, Battery, and Response Speed In Part 8, we shared how we caught hallucinations and improved response quality using 'AI SuperVision'. While making the model smarter and more accurate is a huge milestone, running it in a real-world smartphone environment (like the Galaxy S25 FE) forces us to confront harsh physical walls: Thermal management, Battery consumption, and Latency limits. Unlike the limitless resources of cloud data centers, a mob
Feb 27
![[On-Device AI Chatbot] Part 8: Catching Hallucinations: Analyzing SuperVision Test Results](https://static.wixstatic.com/media/2ea07e_69fba1e933354148a97a50bbfb2f2dcb~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_69fba1e933354148a97a50bbfb2f2dcb~mv2.webp)
![[On-Device AI Chatbot] Part 8: Catching Hallucinations: Analyzing SuperVision Test Results](https://static.wixstatic.com/media/2ea07e_69fba1e933354148a97a50bbfb2f2dcb~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_69fba1e933354148a97a50bbfb2f2dcb~mv2.webp)
[On-Device AI Chatbot] Part 8: Catching Hallucinations: Analyzing SuperVision Test Results
Catching Hallucinations Analyzing SuperVision Test Results In Part 7, we built an automated testing pipeline that bridged our on-device chatbot app inside a smartphone with the AI SuperVision server on a PC. This enabled an end-to-end flow from prompt injection and answer extraction to automated grading. We finally had an environment capable of running dozens of test cases automatically. So, what kind of report card did our on-device SLM (Gemma-2B based) receive from these
Feb 24
![[On-Device AI Chatbot] Part 7: Building SuperVision: An Automated Chatbot Testing Pipeline](https://static.wixstatic.com/media/2ea07e_22b8a8781b1743cb8aaa018b782ab4da~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_22b8a8781b1743cb8aaa018b782ab4da~mv2.webp)
![[On-Device AI Chatbot] Part 7: Building SuperVision: An Automated Chatbot Testing Pipeline](https://static.wixstatic.com/media/2ea07e_22b8a8781b1743cb8aaa018b782ab4da~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_22b8a8781b1743cb8aaa018b782ab4da~mv2.webp)
[On-Device AI Chatbot] Part 7: Building SuperVision: An Automated Chatbot Testing Pipeline
Building SuperVision An Automated Chatbot Testing Pipeline In Part 6, we explained the background of introducing Testworks' 'AI SuperVision' tool to objectively evaluate the chronic hallucination issues inherent in generative AI. However, to actually apply this tool to our project, we had to overcome a significant technical barrier. Our LLM chatbot operates completely offline "On-device" (inside a smartphone), whereas the AI SuperVision system evaluating it exists in a "PC
Feb 23
![[On-Device AI Chatbot] Part 6: How to Verify AI Quality? (Introduction to SuperVision)](https://static.wixstatic.com/media/2ea07e_38184c3eec5940288ae0fcc2e73f6e2d~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_38184c3eec5940288ae0fcc2e73f6e2d~mv2.webp)
![[On-Device AI Chatbot] Part 6: How to Verify AI Quality? (Introduction to SuperVision)](https://static.wixstatic.com/media/2ea07e_38184c3eec5940288ae0fcc2e73f6e2d~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_38184c3eec5940288ae0fcc2e73f6e2d~mv2.webp)
[On-Device AI Chatbot] Part 6: How to Verify AI Quality? (Introduction to SuperVision)
How to Verify AI Quality? Introduction to SuperVision In Part 5, we explored how to inject our company's proprietary knowledge into the on-device chatbot using Local RAG (Retrieval-Augmented Generation) and Multi-Context Switching. However, equipping the chatbot with knowledge does not immediately solve all problems. "How can we be absolutely sure that this chatbot isn't fabricating answers and is truthfully speaking only about what is in the provided documents?" In Part 6,
Feb 21
![[On-Device AI Chatbot] Part 4: The Ears and Mouth of a Chatbot: On-Device STT/TTS Integration](https://static.wixstatic.com/media/2ea07e_f9b2f825229d4e4b8e86be78ac4fd73b~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_f9b2f825229d4e4b8e86be78ac4fd73b~mv2.webp)
![[On-Device AI Chatbot] Part 4: The Ears and Mouth of a Chatbot: On-Device STT/TTS Integration](https://static.wixstatic.com/media/2ea07e_f9b2f825229d4e4b8e86be78ac4fd73b~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_f9b2f825229d4e4b8e86be78ac4fd73b~mv2.webp)
[On-Device AI Chatbot] Part 4: The Ears and Mouth of a Chatbot: On-Device STT/TTS Integration
The Ears and Mouth of a Chatbot On-Device STT/TTS Integration In Part 3, we explored the optimization process of compressing a massive language model to fit the constrained resources of a smartphone and boosting inference speed using the mobile NPU. Now that we have successfully embedded a fast and smart "brain" inside the device, it is time to give our chatbot the "ears and mouth" it needs to interact naturally with users. In a mobile environment, typing out long texts eve
Feb 20
![[On-Device AI Chatbot] Part 5: A Chatbot That Understands Context: Implementing Local RAG and Multi-Context Switching](https://static.wixstatic.com/media/2ea07e_42172a5ac3454535a81160a2408d0b5b~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_42172a5ac3454535a81160a2408d0b5b~mv2.webp)
![[On-Device AI Chatbot] Part 5: A Chatbot That Understands Context: Implementing Local RAG and Multi-Context Switching](https://static.wixstatic.com/media/2ea07e_42172a5ac3454535a81160a2408d0b5b~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_42172a5ac3454535a81160a2408d0b5b~mv2.webp)
[On-Device AI Chatbot] Part 5: A Chatbot That Understands Context: Implementing Local RAG and Multi-Context Switching
A Chatbot That Understands Context Implementing Local RAG and Multi-Context Switching In Part 4, we gave our chatbot "eyes, ears, and a mouth" by integrating on-device STT and TTS. However, no matter how well a chatbot listens and speaks, it is only half-useful as a business assistant if it doesn't know your specific "domain knowledge"—like internal company regulations or specific product manuals. Because Small Language Models (SLMs) are compact, they do not perform as well
Feb 19
![[On-Device AI Chatbot] Part 3: Core Technologies of Mobile AI: Quantization and NPU Optimization](https://static.wixstatic.com/media/2ea07e_08ed983f9efb45fe9129e06967a91163~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_08ed983f9efb45fe9129e06967a91163~mv2.webp)
![[On-Device AI Chatbot] Part 3: Core Technologies of Mobile AI: Quantization and NPU Optimization](https://static.wixstatic.com/media/2ea07e_08ed983f9efb45fe9129e06967a91163~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_08ed983f9efb45fe9129e06967a91163~mv2.webp)
[On-Device AI Chatbot] Part 3: Core Technologies of Mobile AI: Quantization and NPU Optimization
Core Technologies of Mobile AI Quantization and NPU Optimization In Part 2, we discussed our selection of Gemma-2B as the ideal Small Language Model (SLM) for our project and shared our experiences benchmarking CPU and GPU performance in a constrained smartphone environment. However, the initial tests revealed significant challenges: noticeable latency delays and out-of-memory errors. To run LLMs in real-time on a mobile device held in the palm of your hand—not on a data ce
Feb 18
![[On-Device AI Chatbot] Part 2: Giant Language Models in the Palm of Your Hand: Mobile SLM Selection Strategy](https://static.wixstatic.com/media/2ea07e_7ef19534e8cc4690850ed424d904dee6~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_7ef19534e8cc4690850ed424d904dee6~mv2.webp)
![[On-Device AI Chatbot] Part 2: Giant Language Models in the Palm of Your Hand: Mobile SLM Selection Strategy](https://static.wixstatic.com/media/2ea07e_7ef19534e8cc4690850ed424d904dee6~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_7ef19534e8cc4690850ed424d904dee6~mv2.webp)
[On-Device AI Chatbot] Part 2: Giant Language Models in the Palm of Your Hand: Mobile SLM Selection Strategy
Giant Language Models in the Palm of Your Hand Mobile SLM Selection Strategy In Part 1, we explored how "On-Device AI" is becoming an essential paradigm for solving cloud cost and data security issues. But how can we fit massive Large Language Models (LLMs) with tens or hundreds of billions of parameters—which typically run on massive GPU racks in data centers—into a small smartphone? The answer lies in Small Language Models (SLMs) . In Part 2, we will compare the most nota
Feb 17
![[On-Device AI Chatbot] Part 1: Why "On-Device AI" Now? (Overview)](https://static.wixstatic.com/media/2ea07e_fe141ac84a2c46b8b5daf9987efc1ea7~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_fe141ac84a2c46b8b5daf9987efc1ea7~mv2.webp)
![[On-Device AI Chatbot] Part 1: Why "On-Device AI" Now? (Overview)](https://static.wixstatic.com/media/2ea07e_fe141ac84a2c46b8b5daf9987efc1ea7~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_fe141ac84a2c46b8b5daf9987efc1ea7~mv2.webp)
[On-Device AI Chatbot] Part 1: Why "On-Device AI" Now? (Overview)
Why "On-Device AI" Now? Over the past few years, generative AI, led by services like ChatGPT, has revolutionized our daily lives and workflows. However, behind these powerful AI services lies a common limitation: cloud dependency . The standard architecture—where user queries are sent to cloud servers and the computed results from massive data centers are sent back—inevitably introduces risks such as data privacy breaches, network latency, and exorbitant server maintenance c
Feb 16
![[AX Pro] ⑤ Learning to Walk with AI](https://static.wixstatic.com/media/2ea07e_bc6f4a1789f74859bc1fc9663f9eb266~mv2.jpg/v1/fill/w_333,h_250,fp_0.50_0.50,q_30,blur_30,enc_avif,quality_auto/2ea07e_bc6f4a1789f74859bc1fc9663f9eb266~mv2.webp)
![[AX Pro] ⑤ Learning to Walk with AI](https://static.wixstatic.com/media/2ea07e_bc6f4a1789f74859bc1fc9663f9eb266~mv2.jpg/v1/fill/w_300,h_225,fp_0.50_0.50,q_90,enc_avif,quality_auto/2ea07e_bc6f4a1789f74859bc1fc9663f9eb266~mv2.webp)
[AX Pro] ⑤ Learning to Walk with AI
Not Working Alone, But Working 'Together' Through the past four posts, I’ve shared my raw experience with AX Pro, from deployment to real-time operations. As a PO and brand owner, I’ve used countless tools, but few focus as deeply on the essence of 'Human-AI Collaboration' as AX Pro. I didn't just hire a machine that answers questions; I gained a 'Growth-oriented Teammate' that understands our philosophy and evolves under expert guidance. AX Pro main page Three Decisive Mo
Feb 3
![[AX Pro] ④ Operational Evolution: Managing AI with Data, Not Guesswork](https://static.wixstatic.com/media/2ea07e_84e6a80b17474c929a70570cd9ba6b95~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_84e6a80b17474c929a70570cd9ba6b95~mv2.webp)
![[AX Pro] ④ Operational Evolution: Managing AI with Data, Not Guesswork](https://static.wixstatic.com/media/2ea07e_84e6a80b17474c929a70570cd9ba6b95~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_84e6a80b17474c929a70570cd9ba6b95~mv2.webp)
[AX Pro] ④ Operational Evolution: Managing AI with Data, Not Guesswork
Manage AI with metrics, not intuition. AX Pro delivers real-time performance tracking, security with PI Filters, and scalable enterprise AI operations.
Feb 3
![[AX Pro] ③ Taming the AI: "My Chatbot Is Growing Up" (Operations & Feedback)](https://static.wixstatic.com/media/2ea07e_8a73f9acd71242f98ad2f68016dfe489~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_8a73f9acd71242f98ad2f68016dfe489~mv2.webp)
![[AX Pro] ③ Taming the AI: "My Chatbot Is Growing Up" (Operations & Feedback)](https://static.wixstatic.com/media/2ea07e_8a73f9acd71242f98ad2f68016dfe489~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_8a73f9acd71242f98ad2f68016dfe489~mv2.webp)
[AX Pro] ③ Taming the AI: "My Chatbot Is Growing Up" (Operations & Feedback)
Managing the AI's "Puberty" with the Dashboard Deploying the AI isn't the finish line—it's just the beginning. When I first saw the 'Performance Radar' on AX Pro’s main screen, it felt like receiving my son Luke’s report card. Pentagon Metrics : Six key indicators—including Relevance, Accuracy, and Toxicity—are analyzed in real-time and displayed in a radar chart. Overall Score : Our 'Samsung CES 2026' chatbot currently sits at 89 ! A solid grade, but as a PO, I won’t stop
Feb 3
![[AX Pro] ② The 5-Minute Miracle: Building 'Our Own AI' Without a Single Line of Code](https://static.wixstatic.com/media/2ea07e_f459163c86554a718d17ef3bdbd2baa0~mv2.jpg/v1/fill/w_333,h_250,fp_0.50_0.50,q_30,blur_30,enc_avif,quality_auto/2ea07e_f459163c86554a718d17ef3bdbd2baa0~mv2.webp)
![[AX Pro] ② The 5-Minute Miracle: Building 'Our Own AI' Without a Single Line of Code](https://static.wixstatic.com/media/2ea07e_f459163c86554a718d17ef3bdbd2baa0~mv2.jpg/v1/fill/w_300,h_225,fp_0.50_0.50,q_90,enc_avif,quality_auto/2ea07e_f459163c86554a718d17ef3bdbd2baa0~mv2.webp)
[AX Pro] ② The 5-Minute Miracle: Building 'Our Own AI' Without a Single Line of Code
Developers? No, I did it all by myself! Usually, adopting an enterprise solution means drafting a request to the dev team, discussing server specs, and spending a month reading complex API docs. However, true to its name—"Expert Augmentation"—AX Pro has lowered the barrier so that experts can handle the tools directly. From the 'Create Account' screen to hitting 'Create New Group' for my first project, the entire process took less time than it takes for a cup of coffee to c
Feb 3
![[AX Pro] ① AI Taking My Job? Please, Take It Already!](https://static.wixstatic.com/media/2ea07e_4c5daba2872e496a9a2dd81c0b7ad362~mv2.png/v1/fill/w_333,h_250,fp_0.50_0.50,q_35,blur_30,enc_avif,quality_auto/2ea07e_4c5daba2872e496a9a2dd81c0b7ad362~mv2.webp)
![[AX Pro] ① AI Taking My Job? Please, Take It Already!](https://static.wixstatic.com/media/2ea07e_4c5daba2872e496a9a2dd81c0b7ad362~mv2.png/v1/fill/w_300,h_225,fp_0.50_0.50,q_95,enc_avif,quality_auto/2ea07e_4c5daba2872e496a9a2dd81c0b7ad362~mv2.webp)
[AX Pro] ① AI Taking My Job? Please, Take It Already!
AX Pro Studio "Is AI really a threat to our careers?" Everywhere you look, it’s all about AI. News headlines warn us that AI will replace human labor, and office hallways are filled with half-joking concerns about when we’ll be packing our bags. But as a PO (Product Owner) and brand owner who never has enough hours in the day, my honest thought was a bit different: "I wish someone—anyone—would just take these repetitive tasks and tedious data sorting off my plate!" Perhaps
Feb 3


AI Supervision 10. The Blueprint for RAG Success: Integrating AI Supervision into Your Architecture
"We built a RAG system, but where exactly does the evaluation tool fit in?" "How do we map the retrieved documents to the actual answer for validation?" The final puzzle piece in LLM service development is Architecture . It’s not just about calling an API; it’s about creating a seamless pipeline that Retrieves, Generates, and Evaluates. In this final article of our series, we present a practical blueprint for integrating AI Supervision into your RAG (Retrieval-Augmented Ge
Jan 20


AI Supervision 9. AI Beyond the Web: Seamless Evaluation with SDKs and Mobile Integration
"Our AI chatbot lives in a mobile app. Do we have to test it on a separate web dashboard?" "Copy-pasting logs from our server to the evaluation tool is tedious." Many AI evaluation tools are stuck in the browser sandbox. However, real users interact with AI in mobile apps, internal messengers like Slack, or complex backend workflows. The gap between the testing environment and the production environment often leads to unexpected bugs. AI Supervision bridges this gap with r
Jan 20


AI Supervision 8. GPT vs. Claude? Stop Guessing: Precision Model Comparison & Trend Analysis
"I tweaked the prompt, but now the answers feel weird." "I want to switch to a cheaper model, but I'm scared the quality will drop." AI development is a constant series of Trade-offs . You have to decide whether to switch models, adjust prompts, or tune RAG settings. However, looking at just the "Average Score" hides the critical details necessary for these decisions. Use AI Supervision 's Detailed Analysis & Comparison features to put your model under a microscope and see
Jan 20


AI Supervision 7. Cut Costs, Boost Speed: Mastering the Real-time Insights Dashboard
"Why is our API bill so high this month?" "The answer quality is great, but it's too slow for users to wait." For AI development teams, the challenges don't end with "Accuracy." As a service approaches commercialization, it hits the realistic barriers of Latency and Operational Cost . Even a high-quality model will fail if it's too expensive to run or too sluggish for the user. Here is how you can use AI Supervision 's Real-time Insights Dashboard to visualize and optimiz
Jan 20


AI Supervision 6. No More 'test_final_v2.xlsx': Mastering Systematic TestSet Management
"Where is the dataset we used for the last evaluation?" "Is the file Dave sent the latest version?" As you develop AI models, evaluation data files tend to scatter across Slack channels and local drives, with filenames evolving into chaos like v1, final, real_final. If your data isn't managed, your evaluation results cannot be trusted. It’s time to ditch the inefficient file-based workflow. Build a centralized TestSet Management System with AI Supervision . Systematic Test
Jan 20
SECURE YOUR BUSINESS TODAY
bottom of page