Pages

Saturday, August 30, 2025

Stochastic AI Agility: addressing cycle of debts

 The launch of ChatGPT in November 2022 has influenced and perhaps even changed the rules of the game in many areas. Although I personally focus on the IT industry, observations suggest that it is affecting the entire industry, including the lives of ordinary people.


The article aims to discuss or address the observed changes at the project management level. Such observations are addressed by rule 17. Stochastic AI Agility. In recent decades, the industry has been intensively trying to implement agile methodologies to iteratively deliver products.


Some

Figure 1. :  Agile process schema connected to the AI-LLMs (blue lines and circles) and human knowledge based on experiences (blue bricks)


The core idea is to have all processes divided into smaller batches that teams try to implement as best as possible or do the best job possible, especially in IT, maybe not only in this regard. During the agile process, teams learn to accurately estimate delivery times, which in my experience is a very difficult task due to many factors. Such factors usually create visible or invisible technical debts that the teams have to deal with, and can lead to various scenarios.


Forbes Technology Council has identified 16 obstacles[1] to a successful software project that affect virtually every chosen development scenario:


1. Poor collaboration between the product and Engineering teams

2. Not managing data integrity

3. Not aligning early on the ‘must-haves’

4. Overlooking nonfunctional requirements

5. An Unintuitive UI

6. Unexpected Complexities

7. Missed Deadlines

8. Understanding the time needed

9. Scope creep

10. An undefined project scope

11. Unclear or undefined client expectations

12. Overlooking speed, security or the UX

13. Security as an afterthought

14. Hyper-focused planning and design

15. Undisciplined backlog grooming

16. Unclear or incomplete product requirements document


The articles suggest to add one additional point which may be called 

17.  Stochastic AI Agility


The name “Stochastic AI Agility”  suggests that the output of using AI-LLM definitely contributes to the goal, but the impact may not be exactly predicted. The normal distribution may indicate the level of AI-LLM contribution around the most unknown stages, actually areas of the product.

Figure 2.: Agile Kandban schema with each stages connected to the AI-LLMs (blue lines and circles) and human knowledge based on experiences (blue bricks)

The scenarios are later evaluated by management or technical staff with the hope of a better iteration process next time. Yes, we will try to do it better next time and maybe differently.

Figure 3.: Agile Scrum framework schema with each stages connected to the AI-LLMs (blue lines and circles) and human knowledge based on experiences (blue bricks)


It is a good idea next time. With the development and expansion of the Large Lange Models (LLM) solutions, at basically every stage of development or evolution of anything, things can seem a bit biased, even if it is not visible at first glance.

Figure 4.: Waterfall schema  with each stages connected to the AI-LLMs (blue lines and circles) and human knowledge based on experiences (blue bricks)


In my humble opinion, regardless of whether agile (Figure 1,2,3) or non-agile (Figure 4) methodologies are currently used, AI-LLM influences each stage of project development and methodologies.  I suggest future research into agile methodologies, as the use of LLM, Vibe-coding, or managerial decision-making based on LLM advice may lead to unsatisfactory results with broad impact. 


The point 17, Stochasti AI agility, may be a stochastic  process. It may follow a cyclical pattern based on biased recommendations. However, such a fact lowers the probability of success depending on the intensity of the LLM contribution. LLM does not deliver consistent responses, non deterministic. 


Each such inconsistency can lead to debts of varying severity. Although it may not be obvious, each project or goal is ultimately a deterministic process with a level of variance, where small batches, bricks, are built up and these lead to the desired goal. It all depends on the level of complexity we are able to focus on to obtain such bricks, small batches (Figure 6.).


Figure 6.: Every debt and challenge can be broken down into small batches, bricks, to achieve a goal.

PS: Maybe you feel problems or need to rethink and adjust existing approaches, let's think about it together and talk about it.


References:


[1] 16 Obstacles To A Successful Software Project (And How To Avoid Them), https://www.forbes.com/councils/forbestechcouncil/2022/06/21/16-obstacles-to-a-successful-software-project-and-how-to-avoid-them/

Thursday, August 7, 2025

Value of AI knowledge: Does evolution call for deterministic or LLM biased results?

I remember the first time I saw a diagram of a neuron (Img.1.). I was a school kid, and among other hobbies, I was interested in the human brain, it fascinated me. I borrowed a book from the bookstore to find out how it all worked, because these cells are responsible for more than just the functioning of the human species. 

Img.1: Anatomy of multipolar neuron 

Simplification of neuron, perceptron

I took a courses at university focused on neural networks and their applications, basically research, in engineering and control systems. It was a lot of fun because the neuron was simplified into an abstraction presented as a perceptron (Img.2.).

Img.2: Schematic of a perceptron where weights are applied to the inputs, resulting in a weighted sum. The sum is passed to a step function that provides the output.


Using multilayer perceptrons with feedforward learning or back-propagation algorithms, we were able to detect specific patterns that signaled the point of interest. With increasing complexity we faced difficulties in interpreting the results in the deterministic way. Nevertheless, the general result was astonishing.

A breakthrough multi-layer neural network system at massive scale 

Nowadays OpenAI has made a remarkable breakthrough and released the agentic system ChatGPT  in fall  2022.  It was impressive to observe its capabilities of probabilistic evaluation of the next upcoming word.  The technology based on transformer architecture [3] , which may use multilayered perceptron networks behind the scene, was acting like a “real assistant”. It was and is impressive although it is just mathematical calculations on a large scale. It means the true gold of such a system are the calculated weights.

From weights to information and back

Weights required to be properly trained. But such trained weights may contain a noticeable level of entropy [4]. The concept of entropy is used across the disciples and is associated with the level of disorder or randomness of such a system. Such a high level of entropy may be projected to unexpected behaviour. Some research work suggested could be the property of the Large Language Model (LLM) itself.


Why does entropy increase in LLM-based AI systems? LLM models excel at detecting patterns invisible to the naked eye thanks to the concept of layered neural networks and trained weights. The connections, correlations between weights, and how they are updated may be considered non-deterministic. The process can appear to be empirical or stochastic/random. This means that once the weights reach a desired state, it may be challenging to reproduce the exact process. 

Is Boolean algebra still valid?

This reminds me of an operation I learned in my boolean algebra classes at university, material implication (Img.3.)

The material implication applies on two following conditional states x and y  that when x condition is false, the result is always true independently on the condition y.  In other words all is correct. 

Img.3.: Material implication table of 2 conditional states x and y


Perhaps such a state, where everything is correct because the assumption of determinism is false, could be a high-level explanation for the commonly observed LLM state called hallucinations. And as some scientists have already pointed out, Hallucination as the property of the system is difficult to correct. 

Are LLMs a robust repository of knowledge?

The ability of LLM to highlight hidden patterns in text, speech, images, voice has been already mentioned as astonishing. LLM works very well as a translation of a written or spoken word into bytes with a probabilistic level of accuracy identified by an external human observer. The human observer evaluates the received answer/result from the agentic AI system based on LLM. 

The architectural scale of agentic systems may remind someone of, for example, the precisely designed ancient pyramids

The reason why the pyramid is taken into the consideration is the magnitude of the involvement of LLM based systems across the industry. While the architecture may not be seen from a higher perspective as deterministic as that of the Great Pyramids of Giza, its magnitude is evident. These pyramids are considered some of the masterpieces ever built by mankind. Is the use of LLM or Deep learning techniques heading in the same direction?

 

The process of how and with what type of tools the pyramids were built seems to be forever lost to history. It looks similar to how the LLM achieved the most accurate results. The answer lies in the history of weights, or rather how the weights were calculated, which is also lost, but in a very short time frame.


Img.4.: Simplified information storage in architecture pattern, each stone inside the pyramid have deterministically defined structure and position, although fluctuation of material particles may be considered non-deterministic.This is in contrast to a neural network and its storage/extraction of information on the right side of the image. Read color indicates the element to be corrected.

 

What is the knowledge self repair factor ?

Given the robustness and resiliency of the pyramids, the architecture seems to have been proven by time. Any repair can be made at the deepest layer of the system without compromising its stability (Img.4.).  Since LLM AI agent systems are based on multilayer neural networks, which may imply a higher level of entropy, randomness, such a correction, or indeed any correction, may present a challenge with a not-entirely-deterministic outcome. 


The question of repairability of such a complex and highly scalable system may address new discussions. Over the years, mankind has developed a number of enterprise design patterns for business processes that allow performing work with minimal entropy, leading to the desired result with a defined probability, i.e., an error state is considered. Such a process is deterministic and, moreover, repeatable. 

An example is the SAGA pattern (long-running transaction [6]), where the system has the ability to recover from specific states or take action, such state is repeatable.

Is biased information random ?

The intensive development in the field of large language models reveals new insights every day and offers possible directions for improving or mitigating current challenges observed or identified in current implementations. It is possible that as humanity works with increasingly large models, deterministic information sinks deeper into the value of the probability weight. This means that we lose connected points in favor of computed weights. The side effect may occur in disability of effectively promting the question to the system (Img.5.)

Img.5.: A possible consequence of the intense experience of promting with an LLM agents may be damage to cognitive connectivity within the brain due to the removal of connections between stored information.


The trigonometric functions sine and cosine are well known [7] (Img.6.) LLM transformers make extensive use of them to identify the correct word position. It is worth remembering that the entire LLM model still only predicts the next word.

Img.6. Sine and cosine functions plotted from 0 to 2Ï€. 


The use of trigonometric functions has an amazing effect on word order correction. Based on the calculated LLM weights, the correction of dramatically misspelled text is astounding. However, the entire outcome of the agent system may lead to random results because the response is biased by the training data. During current development they may be observed willing to address such questions as biased data may have broader impact.

Prompting conclusion

History has shown the power of deterministically stored information in many different areas. Humanity has trained the brains to increase neuronal connectivity and improve cognitive and other functions through evolution (Img.7.)

Img.7. Random three people who are dedicated to gathering knowledge and understanding through the process of learning, like Mr. Leonardo DaVinci, Tesla and many others

Img.8.: Combining knowledge influenced by the fundamental concept of the gravity projected into the of the pyramid repair process


It is very exciting to watch and contribute to the entire process of evolution in the field of artificial intelligence, whether as an individual or a group. We are constantly discovering many directions from very different perspectives, knowing that our known world must follow a number of fundamental rules, such as gravity, which exist between all entities with mass or energy (Img. 8). The changes in the field of agentic AI systems are very intense and rapid, but the process itself is very narrow and centered around small groups of authors. The scope of impact is very wide. The ability to understand, design or cope with the current state has already predetermined requirements defined by the level of applied mathematics.


The future is not yet written and every moment is important.


Willingness to contribute to the development, among other things, I helped create the JC-AI newsletter [8][9] as part of our Java Champion initiative. Do not hesitate to contact me for future collaboration.



[1] https://en.wikipedia.org/wiki/Neuron

[2] https://en.wikipedia.org/wiki/Perceptron

[3] https://en.wikipedia.org/wiki/Transformer_(deep_learning_architecture)

[4] https://en.wikipedia.org/wiki/Entropy

[5] https://en.wikipedia.org/wiki/Boolean_algebra

[6] https://en.wikipedia.org/wiki/Long-running_transaction

[7] https://en.wikipedia.org/wiki/Sine_and_cosine

[8] https://foojay.io/today/ai-newsletter-1

[9] https://foojay.io/today/jc-ai-newsletter-2


Sunday, August 3, 2025

Stay tuned Artificial Intelligence, not only Java ?

It may seem that nowadays, when the article does not use the abbreviation LLM (Large Language Model) for  AI (Artificial Intelligence), the article has a reduced value, although its message may have had some value.

It may be that the general information flow is overwhelmed with news about the use of artificial intelligence systems. I can offer another perspective, the perspective where the public is eager to understand. Yes, to understand what these artificial intelligence systems are and how they contribute, deterministically contribute, to everyday business or the development of humanity.

In recent years, I have worked on various types of machine learning or knowledge base systems. Some of them used probability theories and advanced mathematics to calculate the most likely outcomes. It all requires continual reading of research articles or modifying current approaches based on newly gained understanding. I feel like field of current Large Language Models based on neural networks and weights requires the same in order to keep understanding what has been achieved and evaluate the current state. 

Knowledge is language independent, but it also requires the right stimulation. I was pleased to see that in our JavaChampion group there is a great need to share knowledge and actively contribute to the process of spreading awareness about all the possibilities not only in the field of LLM, but also in all areas of artificial intelligence in general. Artificial intelligence has much more to offer and it is worth discovering.

Foojay.io - JC-AI-Newsletter Vol. 1

Friday, February 10, 2023

Launch Announcement: How I become an Book Author

"One day, while reading a book, I thought about what it would be like to be an author and write a book myself. This day is here and it is real and here some details of my journey... #java #designpatterns #platform #effectiveness #maintenance #fun #rocks Packt"

  I started working on the book almost a year ago and now it's done! The book is published, all minor tasks are solved and the current state? I'm looking forward to my hardcopies! 

  Now I'm trying to figure out how to properly share how much thoughts I put into creating something that every developer will potentially appreciate on his daily job. Something that could kick her/him back on the trail during being in the local minimum. Maybe not only her/him but also to me just to refresh some points and to keep all knowledge fresh.  I was thinking that maybe even my kids will appreciate it when they grow up, haha.

  Design patterns are a very engaging topic, similar to math. Technologies may change, hardware may change, and we can not stop evolution or time but math remains, similar to the design patterns. They may be adopted differently due to the technology jump but they will be there. 

  The book begins by introducing the Java platform in enough detail to shape the context to understand the value of using patterns. Insights are automatically revealed during usage of presented programming concepts while implementing patterns. 

  I have used neutral examples in the book by using vehicle manufacturing abstractions to drive the reader through the entire book as we all love vehicles. This setup allowed connecting all dots between different design pattern approaches and implementations and to create a flow where the reader may identify himself or herself with the chapter. With all the great Java API’s and all the newly added Java platform enhancements I was inspired  to not use additional frameworks, just pure Java and command-line. I hope the reader appreciates it similarly to me :). In my eyes this allows the reader to stay fully concentrated on the particular topic and apply it across different scenarios. I'll let it upon the reader how successfully I did it. 

  The book contains many standard terms used across the application designers community which makes the book valuable reading material not only for developers but maybe also for project managers to assimilate a similar terminology used across the different types of meetings in different stages of application development. Let's see, it was one of my secret wishes ;)

  Anyhow, after many years of working with multiple languages running on the Java Virtual Machine, my biggest pleasure was always with Java language as the most effective tool to create byte-code. 

  I want to thank my beloved wife, my beautiful kids for giving me energy to step over difficulties and continue my work on this book till the successful end.

It was a great pleasure to work with the Packt team, reviewers  that helped me through this amazing journey. My special thanks goes to Bruno Souza for writing such a beautiful  foreword! 

Thank you guys: Bruno SouzaSonia Chauhan , Sathya MohanPrajakta NaikRohit Kumar SinghWerner Keil and others. 

It was my big pleasure to make this book happen.

Of course I can not forget my peers from the OpenValue family for having some nice discussions with me.