The typical monthly production planning method precipitates several for pharma companies. Only a dynamic approach that ensures flexibility on the shop floor can sustainably ensure reliability and high output.
Tired of meetings? This article discusses why companies should focus on problem-solving productive meetings rather than unproductive “who to blame” reviews.
This article discusses the challenges faced by organizations in sustaining success of transformation initiatives and how to overcome them.
The OEM and tier-1 vendors often face conflict and mistrust, despite needing each other for survival. This article explores solutions for a win-win collaboration.
The automotive new product development (NPD) ecosystem is one of the most complex environments. This article talks about how a finished automobile at the backend has manufacturers toiling away to achieve a complex and audacious end goal.
This article discusses the problems with the basic assumptions of the concept of JIT due to which it failed during the pandemic.
Find out how NPD departments can ensure faster product development and avoid missing planned launch dates
Organizations are abandoning ‘Management by Targets’ for a radical new approach that nurtures both accountability and collaboration
Bajaj Electricals Limited’s transformation story featured in SCM Now magazine – its award winning transformation from a ‘push’ to a ‘pull’ based supply chain and distribution system.
Find out how to protect delivery due dates in manufacturing plants with wandering bottlenecks.
Kiran Kothekar in an investor meet discusses how TOC can help building material companies overcome challenges in the current pandemic crisis .
Covid-19 pandemic represents a novel dilemma – Do you choose a country’s economy or health of its people? What is the way out of this dilemma?
Find out how to overcome confirmation bias in decision making. It is high time that organizations invests in improving the thoughtware for getting more and precise data elements through additional investments in hardware and software.
Find out what creates toxic work culture in an organization?
Read the article to know about how running too many initiatives simultaneously may be counter-productive
Find out why the Balanced scorecard is a weak strategic management tool!
In an interview with Outlook Business, Satyashri Mohanty explains on how Vector team identified the real problem faced at Raymond’s textile division .
With the new addition, the team size of Vector will be 140 consultants .
By implementing “pull” manufacturing and distribution, Raymond is dynamically improving its ability to react quickly to demand. .
Schneider Electric and partner Vector Consulting Group, a supplychain operations firm diagnosed the core issue in software development environments. .
With the evolution of new manufacturing techniques, manufacturing plants became increasingly
Amit Gade’s cheque bounced! Again! When it happened two months back in the month of May, it was dismissed as a one off aberration till it happened again; this was the third in as many months!
In era when accounting was invented, bulk of the cost of a product was truly variable. However in the current era, period expense has become substantial. So, when a significant portion of expenses is artificially allocated to a product, it creates many problems when these variables are used as measurements for decision making.
In Greek Mythology, Procrustes, a bandit, killed his victims by force–fitting them into a standard iron bed. If the victims were shorter than the bed, he would forcefully stretch them; if taller, he would cut off their limbs as per the size of the bed.
This torture technique can be compared to how a predefined standard scale is used to forcefully “size up” work in an environment of high variability.
Sprint-based agile methodology of software development is suffering from the Procrustes phenomenon. This concept’s boundary conditions of applicability are highly limited in the real world. To understand why and how, one needs to go back to the history of software project management processes.
The Dubious History
The history of software projects delivery performance is dubious. Probabilistically speaking, as per the annual Chaos Report, software projects are most likely to fail in terms of cost, quality and timelines. Interestingly, the1 trend has remained the same through the years.
There is now wide spread consensus to declare Waterfall Methodology of software development as a major reason for the poor track record.
Problem with Waterfall Methodology: Process Over Speed?
The waterfall methodology meets two main objectives in software development projects
- Ensure clear accountability in relationship between supplier (one who is developing the software) and the customer (one who will use it)
- Ensure process is highly stable, and that level of rework is minimal
For the first need, freezing requirements before starting the project is an absolute must. Once requirements are frozen, the accountability for any effort or time overrun can be clearly traced back to either of the stakeholders.
The second need requires one to approach software development like one would approach development of a new physical product. Developing a new physical product requires the design to be complete before start of the actual building. Allowing the design to remain fluid during implementation leads to uncontrolled rework and interruptions.
Borrowing from these principles, a waterfall methodology requires one to move the project into following phases sequentially with clear “gates”, marking completion of each phase
Strict change management processes are to be put in place as a “deterrent” against any rethink late in the process. As a result, one of the overheads of this methodology is lot of documentation and sign-offs.
However, when the customer-testing phase comes at the fag end of the entire project, “hindsight” clarity issues create severe conflicts between customer and developer on what was “told” initially in the requirement phase and what was “understood”. So, practically speaking, a strict change management process becomes difficult to implement, more so when the relationship between customer and supplier is that of unequals. Resultant effect is scope rework detected too late in the process, leading to uncontrolled delays. It is no surprise that contractual conflicts, delays, effort spiking and associated stress towards the end of the project are chronic to software development projects.
Agile Development: Speed over Process?
In sharp contrast, as the proponents claim, Agile Methodology takes a viewpoint that scope and design cannot be frozen upfront. Hence, it adopts a flexible working approach, which can embrace change much later in the development cycle. Instead of taking a “one perfect project delivery” viewpoint of software development, Agile believes in the principles of delivering usable features, good enough for customers to start using, and then keep improving based on frequent feedback. The idea is to deliver workable software as soon as possible and get feedback for further iterative development. This way of working also takes away the need to freeze requirements and design upfront.
So, in the Agile world, one can move away from sequential working between phases of software development. The team can work as a self-organizing group, which can design, develop, test and deliver workable features in an iterative manner. Time is managed by using concept of sprints (widely used version of Agile called Scrum) – a “time box” of 15/30 days, is used to deliver small batches of usable software features. While the ongoing sprint is frozen, subsequent ones can be changed based on customer feedback.
The “Not So” Silver Bullet
The silver bullet has been found. Or so, it seems. However, no conclusive, transparent and nondebatable study is available to establish the supremacy of the Scrum Methodology
Beyond the data, what is of concern about Scrum in specific are growing voices of practitioners against the methodology. Numerous blogs have been highlighting the increasing stress of developers, the ever-lengthening list of bugs’ backlog and customer dissatisfaction on usability of software. Surprisingly these effects are exactly opposite of what Agile had promised!
As a counter argument, proponents of the methodology have pointed out that implementation gaps, not conceptual gaps, are the primary reason for the woes.
The argument of implementation gaps puts the onus of the problem of poor results on the abilities of the implementer rather than the inventor. (“They couldn’t implement – they were not disciplined!”). This narrative allows some management theories to stay as fads, longer than they really deserve to. So it is important to verbalize the hidden assumptions (or the boundary conditions) of a theory and check its validity in the practical world.
A widely cherished principle of agile is “self-organizing teams over strict processes”. The signoffs/handovers/checks or any kind of formal control for sequential movement as per a process is scoffed at, as self-organizing and iterative development becomes the guiding mantra.
The resultant effect of any agile intervention is the re-organization of a functional structure of design, development and testing into many small self-organizing teams, which are expected to be self-sufficient in terms of skills of design, development and testing.
This assumption of self-sufficiency of small divided teams in terms of adequacy of high-end skills, particularly designing and requirement assessment is highly questionable in vast majority of environments. High attrition rates, with many developers changing jobs every two to three years are hard realities of the industry, which is obsessed with outsourcing and reducing cost of development. The high attrition rate does not impact availability of generic skill resources like expertise in a programming language. But environment specific skills like design and requirement analysis which is only acquired with adequate experience, are never available in full strength as required by the multiple decentralized agile teams. These skills (also called SMEs or subject matter experts) are always part of a miniscule minority in most environments.
So it is highly likely, with skills decentralization after an agile intervention, that many agile teams will fall short of skills, particularly those contributed by experienced SMEs. Even though, each team may have the required numbers to be declared “self-sufficient” as per records.
This leads to a situation where some agile teams depend informally on others, outside of their team, for help. With decentralization, some teams also become ignorant that they need help outside of their structure! Without a formal process to control sequential movement, the first casualty is adequacy of design and requirement assessment. Hence, testing becomes the primary source of feedback on design and development gaps.
Consequently, frequent re-work becomes inevitable. The lines of codes for same functionality are written many times over in an agile environment. At the same time, continuous “flow back” from testing causes frequent interruptions with developers. The problem of interruptions and rework further aggravates when different small agile groups are working independently on different sets of features in parallel; design decisions, at times, become “local” which cause conflicts between teams and create further rework.
The few SMEs or designers in some self-organizing teams become overloaded. These scarce resources are forced to multi-task beyond their capacity; they have to not just work within their team but also get into issue management of other teams. This further deteriorates the productivity of the system as everyone waits for feedback from heavily loaded resources. End result is an uncontrolled development process.
Agile has an antidote for this inherent chaotic working approach – sprint deadlines. Like Procrustes’s standard iron bed, the sprint cycle deadline force-fits open work fronts into closures. Effort spiking close to sprint end pushes everyone to somehow drive closures and complete the sprint. But the efforts spiking compels compromises – some testing is skipped or some bugs are kept aside, some scope is set aside for later sprints – the same way Procrustes force-fitted human bodies into beds. As almost all resources in the team become extremely busy towards a sprint closure, any planning activity of subsequent sprint is put off to the start of the next sprint.
Consequently, in every sprint, initial time is used up in planning work. Actual development and testing happens towards second half of a sprint cycle, thus building pressure of time, followed by skips and misses. Testing resources stay less utilized during beginning of a sprint and then get overloaded at the end, where all planned features land up together. Sprint end rush and compromises becomes inevitable! The damaging effect of “forcing” a standard sprint cycle is not just seen in execution but also in planning. The arbitrary “cut” of deliverables based on a standard time, at times, lead sprint managers to size out work, which may not be usable at the end of the sprint.
When execution is chaotic, and features coming out from a sprint cycle are not “usable”, the predictability for end customer is lost, particularly for those who might want a firm release schedule for a set of usable features to plan their own projects.
So, in many environments, it is not uncommon to find a hidden waterfall methodology-based project management imposed upon agile working. The resultant effect of this dual environment is a situation, which imbibes the worst of both solutions in terms of productivity losses.
The Hybrid Solution
Typically, an overall project plan is created for “final release” for customer commitment. The longer-term plan is broken into “sprints” with every sprint having its own predefined scope plan. The commitment to end customer is the final release. Sprints turn into typical milestones of an overall project schedule.
When many series of consequent sprints are pre-defined with scope, flexibility to re-scope or repeat a sprint based on feedback of previous sprints becomes nearly impossible. So after every sprint, the pressure is to work on scope of next sprint rather than to deal with feedback/issues of the previous one.
This creates an ever-increasing bug list and leads to massive effort spiking towards the end of the “final release”. Some companies leave aside a time box of “final sprint” before final release to clear all the old mess of previous sprints. So in the real sense, customers don’t get anything usable till the final release.
At times, even the “buffer” of one sprint is not enough to make up for all the past sins. Less available time, pressure to deliver and overload of pending work is a perfect recipe for more compromises. Bad quality products are likely to be delivered to customers who report the bugs back to the development team. Pressure to expand the team to tackle the huge list of bugs of previous releases while developing for new releases increases the cost of development. Stress, poor productivity, delays and bad quality cannot be a rare phenomenon in such an environment.
At the same time, when an overall release project is super-imposed on agile cycles, the typical initial waste of time and development capacity spent in assessing the entire project efforts, scope and plan, negotiations becomes a reality. Software development companies lose out from all sides.
Agile Vs. Waterfall: Solution of Extremes?
One the assumptions of waterfall method, which was questioned by Agile, is the ability of customers to clearly specify requirements upfront. Many times, a requirement becomes clear only when customers actually see or start using something. At the same time, the need for perfection or the “best” can make the requirements phase go out of control. Hence, Agile did away with the overhead of trying to be perfect in design and requirements gathering.
But at same time, Agile implementations also suffer from bane of inadequate designs. What we have now is a more generic conflict underlying the software development process.
- On one hand, we should freeze designs and requirements before implementation because this approach ensures rework is under control and hence, time is under control.
- On the other hand, we should not freeze design and requirements because it is impossible to freeze them. Users get clarity on what they actually want after using or visually observing the product.
The only way out of the above conflict is to understand that there are clearly two types of rework which gets generated in any new product development process.
Type A (Foresight-led rework): They are misses and skips made under pressure of time and lack of synchronization between team members. Some people in the team know about these inputs. Due to pressure of time, these inputs are not incorporated in time; this leads to rework later in the process. These errors cause frustration among team members for missing the obvious. Using testing resources to get feedback on Type A errors is a criminal waste of time and capacity.
Type B (Hindsight-led rework): This is rework created due to new knowledge generated after an event of testing or clear visualization. It is in a way value adding, because new knowledge is generated in the process. Prototypes generate Type 2 value-adding rework. Type B rework is inevitable and no amount of thought experiment can help one visualize the Type B errors upfront. The only way out is to devise a process to generate Type B errors early in the development cycle.
Waterfall Method implicitly assumes that all errors are Type A and hence should be prevented. This school of thought has inspired the manufacturing slogan “First Time Right”. Process rigidity with checks and “no-go” gates becomes the key to avoid Type A lead errors. Agile assumes all errors are Type B, hence it considers one-time handovers between phases as waste of time. Hence agile propagates frequent back and forth between all phases.
Learning from History of Automobile Manufacturing
Interestingly, this “Agile” way is exactly how automobiles used to be manufactured in late 18th century. It was called the craft manufacturing method. A small group of mixed skill resource groups were each given a car to be assembled from start till end. The cross-functional team had skills of design, fitting and even machine operations. The not-so-perfect parts could be filed and adjusted to fit with each other, and eventually a vehicle would come out with arduous rework. This way of car manufacturing was indeed very expensive and only the rich could afford it in that era.
When Henry Ford presented the assembly line concept with clear division of labor, car assembly was transformed. Clear sequential working with different resources specializing in different types of the assembly task, defined the assembly lines of Ford Motors. The cars moved between stations (instead of the workers moving) for continuous flow of output. The specialization on one type of task, along with avoidance of switching losses (of craft manufacturing) between different types of skill work enabled a productivity jump of close to five times.
The revolutionary idea dramatically reduced the cost of making a car. The masses could afford a car because of the enhanced productivity. In a way, the car industry gave up the “agile methodology” of car making and achieved giant strides in productivity.
However, an assembly line working approach requires two critical conditions to be successful
- Flow backs are negligible. Work moves forward in one direction. If nature of work is such that it frequently comes back to previous workstations for rework, then it can create chaos with uncontrolled line stoppages and reduced output. (In such an environment, craft manufacturing, or the agile method, is a better way). In case of Ford , assembly line working was enabled by standardization of parts, which could fit any car, without the need to file and adjust ill-fitting parts for a specific car, thus preventing the flow backs
- Transfer batches should be small (ideally a single piece flow). Transferring of big batches not only expands lead-time in every stage but also leads to late detection of quality issues.
Rapid Feature Flow Model
Let us examine how we can get above two critical conditions to resolve the inherent conflict of software development process.
1. Minimizing Flow Backs
Organizing resource groups into requirements assessment team, design team, development and testing group help create required division of labor for assembly line working. This allows the limited expert resources to be centralized in the requirements and the design groups. This grouping also helps in proper utilization of their capacity in type of work, which requires their expertise. However, without clear definition of what constitutes end of requirement and end of design, the assembly line type sequential working between resource groups will never materialize due to flow backs. It is also important to put in place a process of check to mark completions – this is imperative to break the bad habit of proceeding without completing the phases.
This phase should focus not only defining what should be done in a feature but also defining what will not be given. It is also important to begin with the end in mind. The inputs of all test cases to be used in the final validation should be defined. At times, visual prototypes can also help customers visualize what they really want.
A high-level dependency understanding at an architectural level is usually done in many environments by using the first few sprints, (called the foundational sprints). Dependencies which are neglected are in the subsequent functional sprints where features are delivered.
Design stage should focus on unearthing detailed dependencies in terms of interaction with different functionalities. Identifying verification test cases upfront in this stage makes sure that the gap related to any kind of dependencies which can lead to regression are minimized. In the absence of proper design, lot of capacity wastage can happen later in verification testing as fixing one issue might lead to breaking of other things, creating a vicious loop leading to unpredictability in flow.
The gates of design and requirements along with principle of “beginning with end in mind” should help in preventing type A errors and, unearthing type B errors early in the process.
(Interestingly, engineering of physical products focuses a lot on understanding dependencies as the cost of rework during product building phase is high. For example, if the design of beams needs to accommodate the load bearing capacity of a moving crane in a factory-building project, it is important such design dependencies are thrashed in the engineering phase. If the same dependencies are detected in execution for physical products, the huge physical waste of material becomes visible. However, in the case of software projects, wastages for bad design are never visible; a code going wrong can be re-written in the virtual world after it is detected in testing. No rejection of material or dismantling of build structure happens in the software development world. This inherent advantage of software development is also a bane. When testing is used for facilitating good design, the rework and loss of productivity is colossal, even if there is no loss of physical material).
2. Minimizing the Transfer Batch
Just having an assembly line with delivery coming out after a period of six months or a year is also undesirable. When different customers require different features at different points of time, having one delivery date for a large set of features is of no use to anyone. It is unreasonable to expect the customer to wait till all other features are complete. So, moving individual features through the assembly line in order of priority (sequence in which features are introduced) ensures better flow and a short lead-time in delivering workable features that individual customers want.
The way to make a transfer batch of one feature is to institute a WIP (work in process) limit control in the different resource group types. The work in process control implies that a new feature will be introduced to the resource group only when one under progress is complete as per the criteria set for handovers. The rule of “one in when one is out” ensures a single piece transfer batch flow at a feature level. (Due care should be taken to define features in a way that they are fairly independent entities).
WIP control rule implies that tasks cannot be prescheduled or cut out using the “knife” of a time-boxed sprint. The rule of “one in when one is out” will ensure output of features in a staggered manner and that load on testing is levelled out. (Instead of wave of lumpy work for testing in a sprint-based scheduling.) The staggered flow prevents temporary bottlenecks in the system.
The WIP control also aids transparency of queues, which in turn helps in better capacity planning between resource groups. Skill group sizes can be adjusted time to time to manage any temporary elongation of queue. (Change in complexity of features introduced in the system or loss of resources in some groups can create temporarily lengthen queues).
The above principles will help in rapid flow of features from this system. The rapid delivery of individual deployable features also helps promote “a-version-at-a-time” thinking approach in requirements phase to keep it under control. Customers (or marketers of the software) can keep improving the features by initiating work for the next versions.
The suggested model of working, however, has the following drawbacks
- The rule of WIP control makes scheduling of tasks redundant. In the absence of a time schedule, predictability is lost.
- It is in direct conflict with typical sales planning process of a product company that works on annual or bi-annual software release plans.
The above system exposes queues in the system and helps differentiate touch time (actual work time) from waiting queues. If the features’ content can be standardized as multiples of a unit work package, then it is possible to model the expected arrival time to provide the desired predictability. When flow improves, with drastic reduction in interruptions, priority changes and rework, the system predictability goes up many folds.
But it is important to use the scale of time to provide continuous predictability and not use it as a tool to drive “commitment” in execution or use as a scheduling decision. It is a fallacy that schedules, and commitment around schedules, prevent time wastage. In reality, wastage of time is prevented by daily process of issue resolution and close handholding by flow supervisors in respective groups.
The decoupling of prediction and commitment can be best understood with the analogy of driving using Google Maps. Google Maps offer expected time of arrival looking at queues on road. But it is only a predictive tool, and the driver cannot commit on the time claimed by Google. What he can do effectively is only reduce wastage in every moment of driving. In case of an overall manager of a system, the focus of predictability improvement comes from reducing queues in the system, and managing issues interrupting flow on a daily basis.
In this model, a feature is introduced in the system without trying to cut it out based on some arbitrary sprint cycle definition. The feature size introduced in the system is based on what is deployable.
Dynamic Software Release Planning
The new model mandates that one move away from managing software development as a large monolith project with detailed plan aiming for a single release. Final releases can be done as frequently as one can manage based on the release overheads. Single features or a set of features can be bundled into releases as and when one wants. Instead of committing to a release plan upfront for a year, marketers of the software can decide on release strategy based on dynamics of the market. This provides necessary flexibility to react to competition strategy, and hence change the queue of features in the waiting list to deal with market dynamics.
To summarize, software environments have had a chronic difficulty of not being able to clearly differentiate two independent measuring scales required to manage projects effectively
- Scale of time (to provide predictability of when something will be completed)
- Scale of work to understand size
Both are different; they should not be mixed. Since defining a standard scale for work content has always been a challenge, using time (the sprint cycle duration) as the proxy scale to “knife out” work in planning and in execution is source of all problems in the scrum methodology. Hence, the Procrustes phenomenon becomes unavoidable. The suggested Rapid Feature Flow model offers a way to differentiate the usage of the two scales to escape the wrath of Procrustes.
So what happened to Procrustes? His reign of terror ended with his death at the hands of Theseus, the king of Athens!
Schneider Electric Engineering Workbench, a product from Schneider Electric’s Process Automation business helps in automating the engineering and configuration of the instrumentation and control systems of brown/greenfield projects. Faster engineering helps in reducing lead time to start of operations and thereby improves profitability and reduces the risks of penalties for the Automation vendor. The Agile methodology formerly in place for software development at Schneider Electric Engineering Workbench could not give the desired results. Therefore, the company innovated a flow model based on the core concepts of TOC for the automation software development process which not only created a more harmonious work environment for the development team but also increased productivity by a whopping 71%!
At the end of a year, Schneider reported the following additional benefits
“Innovation”, “Disruption” have been buzzwords for a while. Pundits remind us how Nokia, Kodak didn’t act on time and got disrupted into oblivion.
Any business which manages stock to meet customer needs faces a key challenge in stock management –what is the right inventory?
I think, the best of internalizing new knowledge is by analyzing case studies. So instead of just posting insights, I would like to share a case with you and some questions to ponder at the end of it…
With implementation of Theory of Constraints solution, it is expected that all plant managers align their thinking and decision making to the paradigm of flow rather than local efficiencies.
The framework of Critical Chain Project Management (CCPM) offers a holistic approach for dealing with the Project business. A few of its nuances, however, are far from popular.
Management consulting industry and all IT vendors in enterprise space, promising productivity improvement have a business cycle well aligned with the overall economy.
If one goes by the widely acclaimed Chaos reports released by the Standish Group, software projects seem to have a dubious record. In 1995, the report highlighted that only 16% of the IT projects were completed on time, within budget with original scope.
‘An hour lost on the bottleneck is an hour lost on the entire system; An hour gained on a non-bottleneck is a mirage’
I was addressing an audience of top management of a company recently just before a new assignment was about to kick off.
The logic of CCPM appeals to common sense. The damages of bad multi-tasking and “date driven” behavior are irrefutable.
For any company involved in manufacturing and selling of tangible products, both marketing and supply chain forms an essential aspect of business.
Every literature on project management reminds us that the longest path of dependent tasks determines the total duration of the project.
I once had an opportunity to engage with a company producing and selling custom-built equipment for banks to try and help cut “flab” and steer the company towards improved profitability.
In companies, there is a plethora of improvement projects deployed in any firm at any given time. But, can there be a situation where the very change that people are trying to implement becomes its own biggest hurdle?
The Nano debuted in 2008. It was expected to be a historic event – the launch of the cheapest car ever to be sold on the planet. Auto analysts across the world prophesized the Nano would be a game changer for the Indian auto industry.
Current downturn is putting pressure on companies not just to reduce costs but also to reduce the working capital requirements.
One of the pre-dominant styles of conventional management is management by objectives. On the face of it, the approach appears objective. However, it is not without flaws, as the following case studies highlight.
Almost every manager trying to meet targets, is dealing with obstacles that come his way. He overcomes a few, stumbles over the others. He is identifying constraints and removing them. Is this Theory of Constraints?
Traffic jams are a pain. It disrupts schedules and is a common excuse that people give for reaching late to work. But have you ever wondered what causes these traffic jams? One such day I decided to get out of my car and observe the various events that leads to this event.
Today numerous call centres have been set up to help customers transact, provide transaction updates, act as help lines, answer product related queries, and even handle grievances. The job seems to be over. But Is it?
I am sure most of you must have observed and experienced the extent of disharmony that exists in any organization. Unless sources of disharmony are resolved, situation will not change. Can TOC bring in harmony among people?
The pull system comes into play the moment manufacturing companies or suppliers start with small quantity supplies of new range of SKUs releasing the capital of the distributor. However, to make this reality there are two important paradigms of supply chain that needs to be addressed
People in sales are the most obstinate lot of mankind and it is very difficult to get things implemented in sales. This is the primary belief that most companies hold about the sales function, and is one with quite a bit of truth in it too.
In organizations, leaders must get tasks done through their team members. On the surface, this appears to be an easy thing to do, as leaders have formal authority over their team members (by virtue of organizational hierarchy) and so, can simply issue orders to get things done.
Though Theory of Constraints was originally offered by Eliyahu Goldratt as a manufacturing scheduling method, over the course of his career, Goldratt evolved TOC into a systems methodology which encompasses a wide range of concepts, principles, solutions, tools and approaches
Business ‘Experts’ generally go around departments and ask people for problems, do a gap analysis (with help of some benchmarking data) and then they arrive with solutions, almost one for each problem. Unfortunately, the inter-relationship between the problems is mostly ignored.
Batching issues tend have a profound influence on the flow characteristics of a manufacturing plant and substantial gains can be made by properly understanding underlying principles.
Most managers tend to be “selectors” who screen and judge subordinates based on outcomes! Becoming a coach is harder but a more effective leadership style
According to RBI data, 90% of all transactions in India are made with cash. A little over 80% of this cash became practically useless overnight after Rs1000 and Rs500 currency notes were demonetized on November 8 this year.
Schneider Electric (R&D)/ Vector Consulting Group was chosen as a finalist for the prestigious CSCMP Supply Chain Innovation Award, 2018 .
Skipper Limited turns out to be the only Polymer Products manufacturing company to roll out The Theory of Constraints (TOC) in
Edelweiss as part of their series ‘COFFEE WITH CEO’ the team met Parag Milk Foods’ (Parag) Chairman Mr. Devendra Shah and CFO Mr. Vimal Agarwal to gain insights into its business
Parag Milk Foods announced the results of its new sales pilot after it rolled out TOC Pull distribution solution to provide very high availability at lower inventory while simultaneously improving freshness of products at retail points.
In an interview with Fibre2Fashion, Satyashri Mohanty discusses how Indian textile is losing its “cheap labor” based competitive edge and how exporters have no option but to make a huge paradigm change to regain a compelling advantage in the international markets.
In the recently concluded placement season for 2018, the company has recruited talent from premier institutes like XLRI, IIMs and ISB.
With growing number of industries and business firms, the diversity in the constraints is also growing. Some organization find their way out of the pitfalls where as some grapple with them. Theory of Constraints (TOC) as a reliable solution has benefited largely to these organization. Walking us through the factors associated with the application of TOC and its prevalence, Mr. Satyashri Mohanty, Founder Director of Vector Consulting Group gives us real glimpses of industry.
Pratibha Syntex (Indore), a well-known Indian apparel exporter and a leading manufacturer and exporter of knitted garments and speciality yarns has bagged the ‘Best Global Supplier of the Year Award’ for 2016 by C&A at their Global Supplier Summit in Hongkong.
Finalists Change Management EY with Fannie Mae EY with Monsanto JMW Consultants with Z Energy Corporate Social Responsibility EY with ConAgra Foods KPMG with Ambuja Cement Tata Consultancy Services with Child Maintenance Group Customer Engagement EY with Medtronic Gulland Padfield with Silicon Valley Bank – Asia Tata Consultancy Services with Hilton Grand Vacations Company Developing… Continue reading VCG is 2015 AMCF Awards Finalist
Get in touch
Vector Management Consulting Pvt. Ltd.
10th floor, Thane One, DIL Complex,
Ghodbunder Road, Majiwada,
Thane (West), Maharashtra - 400610, India.
Mr. Hemal Bhuptani