Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are ...Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are various kinds of process models that are used by the software industries for the development of small, medium and long-term software projects, but many of them do not cover risk management. It is quite obvious that the improper selection of the software development process model leads to failure of the software products as it is time bound activity. In the present work, a new software development process model is proposed which covers the risks at any stage of the development of the software product. The model is named a Hemant-Vipin (HV) process model and may be helpful for the software industries for development of the efficient software products and timely delivery at the end of the client. The efficiency of the HV process model is observed by considering various kinds of factors like requirement clarity, user feedback, change agility, predictability, risk identification, practical implementation, customer satisfaction, incremental development, use of ready-made components, quick design, resource organization and many more and found through a case study that the presented approach covers many of parameters in comparison of the existing process models. .展开更多
In recent years,it is the general trend to adopt the standards of international engineering education certification to construct curriculum.“Software Process and Tools”is one of the core courses of Harbin Institute ...In recent years,it is the general trend to adopt the standards of international engineering education certification to construct curriculum.“Software Process and Tools”is one of the core courses of Harbin Institute of Technology’s software engineering undergraduate training program.Focusing on the construction work and practical exploration of the course in the process of reforming the software engineering professional curriculum system,and how to achieve the standards of engineering education certification,This paper makes a review and summary.This paper focuses on the status and the role of the course in the whole curriculum system,as well as project-driven teaching content design and practical teaching methods.And summarizes the experience and results of 3 rounds of teaching practice.展开更多
In order to remain competitive in software industry, software engineers need to continuously improve their proficiency. Personal Software Process (PSPSM) provides a strong concept which helps software engineers inspec...In order to remain competitive in software industry, software engineers need to continuously improve their proficiency. Personal Software Process (PSPSM) provides a strong concept which helps software engineers inspecting and improving themselves. Yet, when being applied on an undergraduate computer engineering course, several complex mathematical calculations from PSP official exercises could encumber the performance of students who do not possess adequate mathematics background. This paper reports a result of implementing PSP course for undergraduate computer engineering students in Chiang Mai University, Thailand.展开更多
The purpose of this paper is to provide the software engineer with tools from the field of manufacturing as an aid to improving software process and product quality. Process involves classical manufacturing methods, s...The purpose of this paper is to provide the software engineer with tools from the field of manufacturing as an aid to improving software process and product quality. Process involves classical manufacturing methods, such as statistical quality control applied to product testing, which is designed to monitor and correct the process when the process yields product quality that fails to meet specifications. Product quality is measured by metrics, such as failure count occurring on software during testing. When the process and product quality are out of control, we show what remedial action to take to bring both the process and product under control. NASA Space Shuttle failure data are used to illustrate the process methods.展开更多
DQPSK modem has been chosen as the modem scheme in many mobile communication systems. A new signal processing technique of π/4-DQPSK modem based on software radio is discussed in this paper. Unlike many other softwar...DQPSK modem has been chosen as the modem scheme in many mobile communication systems. A new signal processing technique of π/4-DQPSK modem based on software radio is discussed in this paper. Unlike many other software radio solutions to the subject, we choose a universal digital radio baseband processor operating as the co-processor of DSP. Only the core algorithms for signal processing are implemented with DSP. Thus the computation burden on DSP is reduced significantly. Compared with the traditional ones, the technique mentioned in this paper is more promising and attractive. It is extremely compact and power-efficient, which is often required by a mobile communication system. The implementation of baseband signal processing for π/4-DQPSK modem on this platform is illustrated in detail. Special emphases are laid on the architecture of the system and the algorithms used in the baseband signal processing. Finally, some experimental results are presented and the performances of the signal processing and compensation algorithms are evaluated through computer simulations.展开更多
Software Process Workshop (SPW 2005) was held in Beijing on May 25-27, 2005. This paper introduces the motivation of organizing such a workshop, as well as its theme and paper gathering and review; and summarizes the ...Software Process Workshop (SPW 2005) was held in Beijing on May 25-27, 2005. This paper introduces the motivation of organizing such a workshop, as well as its theme and paper gathering and review; and summarizes the main content and insights of 11 keynote speeches, 30 regular papers in five sessions of “Process Content”, “Process Tools and Metrics”, “Process Management”, “Process Representation and Analysis”, and “Experience Reports”, 8 software development support tools demonstration, and the ending panel “Where Are We Now? Where Should We Go Next?”.展开更多
According to the characteristic of Team Software Process (TSP), it adopts a hierarchy-based model combined discrete event model with system dynamics model. This model represents TSP as form of two levels, the inner ...According to the characteristic of Team Software Process (TSP), it adopts a hierarchy-based model combined discrete event model with system dynamics model. This model represents TSP as form of two levels, the inner level embodies the continuity of the software process, the outer embodies the software development process by phases, and the structure and principle of the model is explained in detail, then formalization description of the model is offered. At last, an example is presented to demonstrate the simulation process and result. This model can simulate team software process from various angles, supervise and predict the software process. Also it can make the management of software development become more scientific and improve the quality of software.展开更多
Behaviour detection models based on automata have been studied widely. By add- ing edge ε, the local automata are combined into global automata to describe and detect soft- ware behaviour. However, these methods in- ...Behaviour detection models based on automata have been studied widely. By add- ing edge ε, the local automata are combined into global automata to describe and detect soft- ware behaviour. However, these methods in- troduce nondeterminacy, leading to models that are imprecise or inefficient. We present a model of software Behaviour Detection based on Process Algebra and system call (BDPA). In this model, a system call is mapped into an action, and a function is mapped into a process We construct a process expression for each function to describe its behaviour. Without con- strutting automata or introducing nondeter- minacy, we use algebraic properties and algo- rithms to obtain a global process expression by combining the process expressions derived from each function. Behaviour detection rules and methods based on BDPA are determined by equivalence theory. Experiments demon- strate that the BDPA model has better preci- sion and efficiency than traditional methods.展开更多
Due to the randomness and time dependence of the factors affecting software reliability, most software reliability models are treated as stochastic processes, and the non-homogeneous Poisson process(NHPP) is the most ...Due to the randomness and time dependence of the factors affecting software reliability, most software reliability models are treated as stochastic processes, and the non-homogeneous Poisson process(NHPP) is the most used one.However, the failure behavior of software does not follow the NHPP in a statistically rigorous manner, and the pure random method might be not enough to describe the software failure behavior. To solve these problems, this paper proposes a new integrated approach that combines stochastic process and grey system theory to describe the failure behavior of software. A grey NHPP software reliability model is put forward in a discrete form, and a grey-based approach for estimating software reliability under the NHPP is proposed as a nonlinear multi-objective programming problem. Finally, four grey NHPP software reliability models are applied to four real datasets, the dynamic R-square and predictive relative error are calculated. Comparing with the original single NHPP software reliability model, it is found that the modeling using the integrated approach has a higher prediction accuracy of software reliability. Therefore, there is the characteristics of grey uncertain information in the NHPP software reliability models, and exploiting the latent grey uncertain information might lead to more accurate software reliability estimation.展开更多
Software projects influenced by many human factors generate various risks. In order to develop highly quality software, it is important to respond to these risks reasonably and promptly. In addition, it is not easy fo...Software projects influenced by many human factors generate various risks. In order to develop highly quality software, it is important to respond to these risks reasonably and promptly. In addition, it is not easy for project managers to deal with these risks completely. Therefore, it is essential to manage the process quality by promoting activities of process monitoring and design quality assessment. In this paper, we discuss statistical data analysis for actual project management activities in process monitoring and design quality assessment, and analyze the effects for these software process improvement quantitatively by applying the methods of multivariate analysis. Then, we show how process factors affect the management measures of QCD (Quality, Cost, Delivery) by applying the multiple regression analyses to observed process monitoring data. Further, we quantitatively evaluate the effect by performing design quality assessment based on the principal component analysis and the factor analysis. As a result of analysis, we show that the design quality assessment activities are so effective for software process improvement. Further, based on the result of quantitative project assessment, we discuss the usefulness of process monitoring progress assessment by using a software reliability growth model. This result may enable us to give a useful quantitative measure of product release determination.展开更多
The Goel-Okumoto software reliability model, also known as the Exponential Nonhomogeneous Poisson Process,is one of the earliest software reliability models to be proposed. From literature, it is evident that most of ...The Goel-Okumoto software reliability model, also known as the Exponential Nonhomogeneous Poisson Process,is one of the earliest software reliability models to be proposed. From literature, it is evident that most of the study that has been done on the Goel-Okumoto software reliability model is parameter estimation using the MLE method and model fit. It is widely known that predictive analysis is very useful for modifying, debugging and determining when to terminate software development testing process. However, there is a conspicuous absence of literature on both the classical and Bayesian predictive analyses on the model. This paper presents some results about predictive analyses for the Goel-Okumoto software reliability model. Driven by the requirement of highly reliable software used in computers embedded in automotive, mechanical and safety control systems, industrial and quality process control, real-time sensor networks, aircrafts, nuclear reactors among others, we address four issues in single-sample prediction associated closely with software development process. We have adopted Bayesian methods based on non-informative priors to develop explicit solutions to these problems. An example with real data in the form of time between software failures will be used to illustrate the developed methodologies.展开更多
A new approach for assessing and optimizing software project process based on software risk control presented, which evaluates and optimizes software project process from the view of controlling the software project r...A new approach for assessing and optimizing software project process based on software risk control presented, which evaluates and optimizes software project process from the view of controlling the software project risks. A model for optimizing software risk control is given, a discrete optimization algorithm based on dynamic programruing is proposed and an example of using above method to solve a problem is also included in this paper. By improving the old passive post-project control into an active effective preaction, this new method can greatly promote the possibility of success of software projects.展开更多
Because of the inevitable debugging lag,imperfect debugging process is used to replace perfect debugging process in the analysis of software reliability growth model.Considering neither testing-effort nor testing cove...Because of the inevitable debugging lag,imperfect debugging process is used to replace perfect debugging process in the analysis of software reliability growth model.Considering neither testing-effort nor testing coverage can describe software reliability for imperfect debugging completely,by hybridizing testing-effort with testing coverage under imperfect debugging,this paper proposes a new model named GMW-LO-ID.Under the assumption that the number of faults is proportional to the current number of detected faults,this model combines generalized modified Weibull(GMW)testing-effort function with logistic(LO)testing coverage function,and inherits GMW's amazing flexibility and LO's high fitting precision.Furthermore,the fitting accuracy and predictive power are verified by two series of experiments and we can draw a conclusion that our model fits the actual failure data better and predicts the software future behavior better than other ten traditional models,which only consider one or two points of testing-effort,testing coverage and imperfect debugging.展开更多
For a more accurate and comprehensive assessment of the trustworthiness of component-based soft- ware system, the fuzzy analytic hierarchy process is introduced to establish the analysis model. Combine qualitative and...For a more accurate and comprehensive assessment of the trustworthiness of component-based soft- ware system, the fuzzy analytic hierarchy process is introduced to establish the analysis model. Combine qualitative and quantitative analyses, the impacts to overall trustworthiness by the different types of components are distinguished. Considering the coupling relationship between components, dividing the system into several layers from target layer to scheme layer, evaluating the scheme advantages disadvantages by group decision-making, the trustworthiness of a typical J2EE structured component-based software is assessed. The trustworthiness asses model of the software components provides an effective methods of operation.展开更多
Software project outcomes heavily depend on natural language requirements,often causing diverse interpretations and issues like ambiguities and incomplete or faulty requirements.Researchers are exploring machine learn...Software project outcomes heavily depend on natural language requirements,often causing diverse interpretations and issues like ambiguities and incomplete or faulty requirements.Researchers are exploring machine learning to predict software bugs,but a more precise and general approach is needed.Accurate bug prediction is crucial for software evolution and user training,prompting an investigation into deep and ensemble learning methods.However,these studies are not generalized and efficient when extended to other datasets.Therefore,this paper proposed a hybrid approach combining multiple techniques to explore their effectiveness on bug identification problems.The methods involved feature selection,which is used to reduce the dimensionality and redundancy of features and select only the relevant ones;transfer learning is used to train and test the model on different datasets to analyze how much of the learning is passed to other datasets,and ensemble method is utilized to explore the increase in performance upon combining multiple classifiers in a model.Four National Aeronautics and Space Administration(NASA)and four Promise datasets are used in the study,showing an increase in the model’s performance by providing better Area Under the Receiver Operating Characteristic Curve(AUC-ROC)values when different classifiers were combined.It reveals that using an amalgam of techniques such as those used in this study,feature selection,transfer learning,and ensemble methods prove helpful in optimizing the software bug prediction models and providing high-performing,useful end mode.展开更多
Reflection mechanism for reuse software architecture (RMRSA) divides a software architecture into base-level architecture and meta-level architecture logically. Base-level architecture is the ordinary architecture; ...Reflection mechanism for reuse software architecture (RMRSA) divides a software architecture into base-level architecture and meta-level architecture logically. Base-level architecture is the ordinary architecture; meta-level represents and manipulates the reusable meta-information of base-level architecture explicitly. Through reflection, the modification of meta-level architecture will result in the modification of the architecture in base-level. Then we can gain a new base-level architecture design. In this paper, we use π-calculus to define the constituents and their interaction processes of RMRSA, by these definition, we specify the business function in base-level at runtime, and illustrate the reflection mechanism between the base-level architecture and meta-level architecture.展开更多
This research recognizes the limitation and challenges of adaptingand applying Process Mining as a powerful tool and technique in theHypothetical Software Architecture (SA) Evaluation Framework with thefeatures and fa...This research recognizes the limitation and challenges of adaptingand applying Process Mining as a powerful tool and technique in theHypothetical Software Architecture (SA) Evaluation Framework with thefeatures and factors of lightweightness. Process mining deals with the largescalecomplexity of security and performance analysis, which are the goalsof SA evaluation frameworks. As a result of these conjectures, all ProcessMining researches in the realm of SA are thoroughly reviewed, and ninechallenges for Process Mining Adaption are recognized. Process mining isembedded in the framework and to boost the quality of the SA model forfurther analysis, the framework nominates architectural discovery algorithmsFlower, Alpha, Integer Linear Programming (ILP), Heuristic, and Inductiveand compares them vs. twelve quality criteria. Finally, the framework’s testingon three case studies approves the feasibility of applying process mining toarchitectural evaluation. The extraction of the SA model is also done by thebest model discovery algorithm, which is selected by intensive benchmarkingin this research. This research presents case studies of SA in service-oriented,Pipe and Filter, and component-based styles, modeled and simulated byHierarchical Colored Petri Net techniques based on the cases’ documentation.Processminingwithin this framework dealswith the system’s log files obtainedfrom SA simulation. Applying process mining is challenging, especially for aSA evaluation framework, as it has not been done yet. The research recognizesthe problems of process mining adaption to a hypothetical lightweightSA evaluation framework and addresses these problems during the solutiondevelopment.展开更多
In the textile industry,garment manufacturing contains four major pro-cesses containing cutting,sewing,finishing,and packaging.Sewing is the most crucial and intricate section,dealing with a large number of varied oper...In the textile industry,garment manufacturing contains four major pro-cesses containing cutting,sewing,finishing,and packaging.Sewing is the most crucial and intricate section,dealing with a large number of varied operations.A successful sewing process needs to be optimized regarding different factors,including time,sewing equipment,and skilled workers.Assembly lineflow is combined by a set of operations with a particular sequence.The utmost impor-tance of all garment industry is to arrange the workstations to minimize the num-ber of employees in order to produce at the best productive rate with the most reasonable cost,shortest time,and satisfying quality.In most garment factories,the production lines are balanced using the empirical judgment of the line man-agers.For the whole process the data of production time at each step,labor pro-ductivity,proper choices of equipment were always needed to calculate line efficiency.As far as the issue is concerned,there has not been an academically sewing process analyzing software providing adequate data of sewing motions and sewing time as the credible input for the line balancing tasks.Towards this goal,this paper presents the results of research on optimizing academically self-built software to analyze the sewing process of knitted products applied to industrial production using Java programming language on Google tools.The results achieved by the software are not only to analyze sewing products and the technological sewing process,calculate the sewing time on the machine but also analyze the sewing activities of workers into manipulations,movements,and motions to calculate the preparation time for two typical knitted products,namely,Polo-Shirt and T-Shirt with the case studies at General Textile Garment Joint Stock Company Hanoi and Star Fashion Company Limited.展开更多
Software process is a framework for effective and timely delivery of software system. The framework plays a crucial role for software success. However, the development of large-scale software still faces the crisis of...Software process is a framework for effective and timely delivery of software system. The framework plays a crucial role for software success. However, the development of large-scale software still faces the crisis of high risks, low quality, high costs and long cycle time. This paper proposed a three-phase parallel-pipelining software process model for improving speed and productivity, and reducing software costs and risks without sacrificing software quality. In this model, two strategies were presented. One strategy, based on subsystem-cost priority, was used to prevent software development cost wasting and to reduce software complexity as well; the other strategy, used for balancing subsystem complexity, was designed to reduce the software complexity in the later development stages. Moreover, the proposed function-detailed and workload-simplified subsystem pipelining software process model presents much higher parallelity than the concurrent incremental model. Finally, the component-based product line technology not only ensures software quality and further reduces cycle time, software costs, and software risks but also sufficiently and rationally utilizes previous software product resources and enhances the competition ability of software development organizations.展开更多
The Goel-Okumoto software reliability model is one of the earliest attempts to use a non-homogeneous Poisson process to model failure times observed during software test interval. The model is known as exponential NHP...The Goel-Okumoto software reliability model is one of the earliest attempts to use a non-homogeneous Poisson process to model failure times observed during software test interval. The model is known as exponential NHPP model as it describes exponential software failure curve. Parameter estimation, model fit and predictive analyses based on one sample have been conducted on the Goel-Okumoto software reliability model. However, predictive analyses based on two samples have not been conducted on the model. In two-sample prediction, the parameters and characteristics of the first sample are used to analyze and to make predictions for the second sample. This helps in saving time and resources during the software development process. This paper presents some results about predictive analyses for the Goel-Okumoto software reliability model based on two samples. We have addressed three issues in two-sample prediction associated closely with software development testing process. Bayesian methods based on non-informative priors have been adopted to develop solutions to these issues. The developed methodologies have been illustrated by two sets of software failure data simulated from the Goel-Okumoto software reliability model.展开更多
文摘Software Development Life Cycle (SDLC) is one of the major ingredients for the development of efficient software systems within a time frame and low-cost involvement. From the literature, it is evident that there are various kinds of process models that are used by the software industries for the development of small, medium and long-term software projects, but many of them do not cover risk management. It is quite obvious that the improper selection of the software development process model leads to failure of the software products as it is time bound activity. In the present work, a new software development process model is proposed which covers the risks at any stage of the development of the software product. The model is named a Hemant-Vipin (HV) process model and may be helpful for the software industries for development of the efficient software products and timely delivery at the end of the client. The efficiency of the HV process model is observed by considering various kinds of factors like requirement clarity, user feedback, change agility, predictability, risk identification, practical implementation, customer satisfaction, incremental development, use of ready-made components, quick design, resource organization and many more and found through a case study that the presented approach covers many of parameters in comparison of the existing process models. .
文摘In recent years,it is the general trend to adopt the standards of international engineering education certification to construct curriculum.“Software Process and Tools”is one of the core courses of Harbin Institute of Technology’s software engineering undergraduate training program.Focusing on the construction work and practical exploration of the course in the process of reforming the software engineering professional curriculum system,and how to achieve the standards of engineering education certification,This paper makes a review and summary.This paper focuses on the status and the role of the course in the whole curriculum system,as well as project-driven teaching content design and practical teaching methods.And summarizes the experience and results of 3 rounds of teaching practice.
文摘In order to remain competitive in software industry, software engineers need to continuously improve their proficiency. Personal Software Process (PSPSM) provides a strong concept which helps software engineers inspecting and improving themselves. Yet, when being applied on an undergraduate computer engineering course, several complex mathematical calculations from PSP official exercises could encumber the performance of students who do not possess adequate mathematics background. This paper reports a result of implementing PSP course for undergraduate computer engineering students in Chiang Mai University, Thailand.
文摘The purpose of this paper is to provide the software engineer with tools from the field of manufacturing as an aid to improving software process and product quality. Process involves classical manufacturing methods, such as statistical quality control applied to product testing, which is designed to monitor and correct the process when the process yields product quality that fails to meet specifications. Product quality is measured by metrics, such as failure count occurring on software during testing. When the process and product quality are out of control, we show what remedial action to take to bring both the process and product under control. NASA Space Shuttle failure data are used to illustrate the process methods.
文摘DQPSK modem has been chosen as the modem scheme in many mobile communication systems. A new signal processing technique of π/4-DQPSK modem based on software radio is discussed in this paper. Unlike many other software radio solutions to the subject, we choose a universal digital radio baseband processor operating as the co-processor of DSP. Only the core algorithms for signal processing are implemented with DSP. Thus the computation burden on DSP is reduced significantly. Compared with the traditional ones, the technique mentioned in this paper is more promising and attractive. It is extremely compact and power-efficient, which is often required by a mobile communication system. The implementation of baseband signal processing for π/4-DQPSK modem on this platform is illustrated in detail. Special emphases are laid on the architecture of the system and the algorithms used in the baseband signal processing. Finally, some experimental results are presented and the performances of the signal processing and compensation algorithms are evaluated through computer simulations.
文摘Software Process Workshop (SPW 2005) was held in Beijing on May 25-27, 2005. This paper introduces the motivation of organizing such a workshop, as well as its theme and paper gathering and review; and summarizes the main content and insights of 11 keynote speeches, 30 regular papers in five sessions of “Process Content”, “Process Tools and Metrics”, “Process Management”, “Process Representation and Analysis”, and “Experience Reports”, 8 software development support tools demonstration, and the ending panel “Where Are We Now? Where Should We Go Next?”.
基金Supported by the National Defense Basic ResearchFoundation (K1503063165)
文摘According to the characteristic of Team Software Process (TSP), it adopts a hierarchy-based model combined discrete event model with system dynamics model. This model represents TSP as form of two levels, the inner level embodies the continuity of the software process, the outer embodies the software development process by phases, and the structure and principle of the model is explained in detail, then formalization description of the model is offered. At last, an example is presented to demonstrate the simulation process and result. This model can simulate team software process from various angles, supervise and predict the software process. Also it can make the management of software development become more scientific and improve the quality of software.
基金supported by the Fund of National Natural Science Project under Grant No.61272125the Specialized Research Fund for the Doctoral Program of Higher Education under Grant No.20121333110014the Hebei Provincial Natural Science Foundation under Grant No.F2011203234
文摘Behaviour detection models based on automata have been studied widely. By add- ing edge ε, the local automata are combined into global automata to describe and detect soft- ware behaviour. However, these methods in- troduce nondeterminacy, leading to models that are imprecise or inefficient. We present a model of software Behaviour Detection based on Process Algebra and system call (BDPA). In this model, a system call is mapped into an action, and a function is mapped into a process We construct a process expression for each function to describe its behaviour. Without con- strutting automata or introducing nondeter- minacy, we use algebraic properties and algo- rithms to obtain a global process expression by combining the process expressions derived from each function. Behaviour detection rules and methods based on BDPA are determined by equivalence theory. Experiments demon- strate that the BDPA model has better preci- sion and efficiency than traditional methods.
基金supported by the National Natural Science Foundation of China (71671090)the Fundamental Research Funds for the Central Universities (NP2020022)the Qinglan Project of Excellent Youth or Middle-Aged Academic Leaders in Jiangsu Province。
文摘Due to the randomness and time dependence of the factors affecting software reliability, most software reliability models are treated as stochastic processes, and the non-homogeneous Poisson process(NHPP) is the most used one.However, the failure behavior of software does not follow the NHPP in a statistically rigorous manner, and the pure random method might be not enough to describe the software failure behavior. To solve these problems, this paper proposes a new integrated approach that combines stochastic process and grey system theory to describe the failure behavior of software. A grey NHPP software reliability model is put forward in a discrete form, and a grey-based approach for estimating software reliability under the NHPP is proposed as a nonlinear multi-objective programming problem. Finally, four grey NHPP software reliability models are applied to four real datasets, the dynamic R-square and predictive relative error are calculated. Comparing with the original single NHPP software reliability model, it is found that the modeling using the integrated approach has a higher prediction accuracy of software reliability. Therefore, there is the characteristics of grey uncertain information in the NHPP software reliability models, and exploiting the latent grey uncertain information might lead to more accurate software reliability estimation.
文摘Software projects influenced by many human factors generate various risks. In order to develop highly quality software, it is important to respond to these risks reasonably and promptly. In addition, it is not easy for project managers to deal with these risks completely. Therefore, it is essential to manage the process quality by promoting activities of process monitoring and design quality assessment. In this paper, we discuss statistical data analysis for actual project management activities in process monitoring and design quality assessment, and analyze the effects for these software process improvement quantitatively by applying the methods of multivariate analysis. Then, we show how process factors affect the management measures of QCD (Quality, Cost, Delivery) by applying the multiple regression analyses to observed process monitoring data. Further, we quantitatively evaluate the effect by performing design quality assessment based on the principal component analysis and the factor analysis. As a result of analysis, we show that the design quality assessment activities are so effective for software process improvement. Further, based on the result of quantitative project assessment, we discuss the usefulness of process monitoring progress assessment by using a software reliability growth model. This result may enable us to give a useful quantitative measure of product release determination.
文摘The Goel-Okumoto software reliability model, also known as the Exponential Nonhomogeneous Poisson Process,is one of the earliest software reliability models to be proposed. From literature, it is evident that most of the study that has been done on the Goel-Okumoto software reliability model is parameter estimation using the MLE method and model fit. It is widely known that predictive analysis is very useful for modifying, debugging and determining when to terminate software development testing process. However, there is a conspicuous absence of literature on both the classical and Bayesian predictive analyses on the model. This paper presents some results about predictive analyses for the Goel-Okumoto software reliability model. Driven by the requirement of highly reliable software used in computers embedded in automotive, mechanical and safety control systems, industrial and quality process control, real-time sensor networks, aircrafts, nuclear reactors among others, we address four issues in single-sample prediction associated closely with software development process. We have adopted Bayesian methods based on non-informative priors to develop explicit solutions to these problems. An example with real data in the form of time between software failures will be used to illustrate the developed methodologies.
基金Supported bythe Plan of New Technology Projectsin China National Packaging Corporation2005 (05ZBJA011)the Na-tional Natural Science Foundation of China (60373062) National Sci-ence Foundation of Hunan Province(04JJ3052)
文摘A new approach for assessing and optimizing software project process based on software risk control presented, which evaluates and optimizes software project process from the view of controlling the software project risks. A model for optimizing software risk control is given, a discrete optimization algorithm based on dynamic programruing is proposed and an example of using above method to solve a problem is also included in this paper. By improving the old passive post-project control into an active effective preaction, this new method can greatly promote the possibility of success of software projects.
基金supported by the National Natural Science Foundation of China(No.U1433116)the Aviation Science Foundation of China(No.20145752033)
文摘Because of the inevitable debugging lag,imperfect debugging process is used to replace perfect debugging process in the analysis of software reliability growth model.Considering neither testing-effort nor testing coverage can describe software reliability for imperfect debugging completely,by hybridizing testing-effort with testing coverage under imperfect debugging,this paper proposes a new model named GMW-LO-ID.Under the assumption that the number of faults is proportional to the current number of detected faults,this model combines generalized modified Weibull(GMW)testing-effort function with logistic(LO)testing coverage function,and inherits GMW's amazing flexibility and LO's high fitting precision.Furthermore,the fitting accuracy and predictive power are verified by two series of experiments and we can draw a conclusion that our model fits the actual failure data better and predicts the software future behavior better than other ten traditional models,which only consider one or two points of testing-effort,testing coverage and imperfect debugging.
基金Sponsored by the National High Technology Research and Development Program of China ("863"Program) (2009AA01Z433)
文摘For a more accurate and comprehensive assessment of the trustworthiness of component-based soft- ware system, the fuzzy analytic hierarchy process is introduced to establish the analysis model. Combine qualitative and quantitative analyses, the impacts to overall trustworthiness by the different types of components are distinguished. Considering the coupling relationship between components, dividing the system into several layers from target layer to scheme layer, evaluating the scheme advantages disadvantages by group decision-making, the trustworthiness of a typical J2EE structured component-based software is assessed. The trustworthiness asses model of the software components provides an effective methods of operation.
基金This Research is funded by Researchers Supporting Project Number(RSPD2024R947),King Saud University,Riyadh,Saudi Arabia.
文摘Software project outcomes heavily depend on natural language requirements,often causing diverse interpretations and issues like ambiguities and incomplete or faulty requirements.Researchers are exploring machine learning to predict software bugs,but a more precise and general approach is needed.Accurate bug prediction is crucial for software evolution and user training,prompting an investigation into deep and ensemble learning methods.However,these studies are not generalized and efficient when extended to other datasets.Therefore,this paper proposed a hybrid approach combining multiple techniques to explore their effectiveness on bug identification problems.The methods involved feature selection,which is used to reduce the dimensionality and redundancy of features and select only the relevant ones;transfer learning is used to train and test the model on different datasets to analyze how much of the learning is passed to other datasets,and ensemble method is utilized to explore the increase in performance upon combining multiple classifiers in a model.Four National Aeronautics and Space Administration(NASA)and four Promise datasets are used in the study,showing an increase in the model’s performance by providing better Area Under the Receiver Operating Characteristic Curve(AUC-ROC)values when different classifiers were combined.It reveals that using an amalgam of techniques such as those used in this study,feature selection,transfer learning,and ensemble methods prove helpful in optimizing the software bug prediction models and providing high-performing,useful end mode.
基金Supported by the National Natural Science Foundation of China (60473066)Young Outstanding Talent Foundation of Hubei Province,China(2003ABB004)
文摘Reflection mechanism for reuse software architecture (RMRSA) divides a software architecture into base-level architecture and meta-level architecture logically. Base-level architecture is the ordinary architecture; meta-level represents and manipulates the reusable meta-information of base-level architecture explicitly. Through reflection, the modification of meta-level architecture will result in the modification of the architecture in base-level. Then we can gain a new base-level architecture design. In this paper, we use π-calculus to define the constituents and their interaction processes of RMRSA, by these definition, we specify the business function in base-level at runtime, and illustrate the reflection mechanism between the base-level architecture and meta-level architecture.
基金This paper is supported by Research Grant Number:PP-FTSM-2022.
文摘This research recognizes the limitation and challenges of adaptingand applying Process Mining as a powerful tool and technique in theHypothetical Software Architecture (SA) Evaluation Framework with thefeatures and factors of lightweightness. Process mining deals with the largescalecomplexity of security and performance analysis, which are the goalsof SA evaluation frameworks. As a result of these conjectures, all ProcessMining researches in the realm of SA are thoroughly reviewed, and ninechallenges for Process Mining Adaption are recognized. Process mining isembedded in the framework and to boost the quality of the SA model forfurther analysis, the framework nominates architectural discovery algorithmsFlower, Alpha, Integer Linear Programming (ILP), Heuristic, and Inductiveand compares them vs. twelve quality criteria. Finally, the framework’s testingon three case studies approves the feasibility of applying process mining toarchitectural evaluation. The extraction of the SA model is also done by thebest model discovery algorithm, which is selected by intensive benchmarkingin this research. This research presents case studies of SA in service-oriented,Pipe and Filter, and component-based styles, modeled and simulated byHierarchical Colored Petri Net techniques based on the cases’ documentation.Processminingwithin this framework dealswith the system’s log files obtainedfrom SA simulation. Applying process mining is challenging, especially for aSA evaluation framework, as it has not been done yet. The research recognizesthe problems of process mining adaption to a hypothetical lightweightSA evaluation framework and addresses these problems during the solutiondevelopment.
基金This study was carried out within the framework of the topic Science and Technology 01C–02/04–2019–3.
文摘In the textile industry,garment manufacturing contains four major pro-cesses containing cutting,sewing,finishing,and packaging.Sewing is the most crucial and intricate section,dealing with a large number of varied operations.A successful sewing process needs to be optimized regarding different factors,including time,sewing equipment,and skilled workers.Assembly lineflow is combined by a set of operations with a particular sequence.The utmost impor-tance of all garment industry is to arrange the workstations to minimize the num-ber of employees in order to produce at the best productive rate with the most reasonable cost,shortest time,and satisfying quality.In most garment factories,the production lines are balanced using the empirical judgment of the line man-agers.For the whole process the data of production time at each step,labor pro-ductivity,proper choices of equipment were always needed to calculate line efficiency.As far as the issue is concerned,there has not been an academically sewing process analyzing software providing adequate data of sewing motions and sewing time as the credible input for the line balancing tasks.Towards this goal,this paper presents the results of research on optimizing academically self-built software to analyze the sewing process of knitted products applied to industrial production using Java programming language on Google tools.The results achieved by the software are not only to analyze sewing products and the technological sewing process,calculate the sewing time on the machine but also analyze the sewing activities of workers into manipulations,movements,and motions to calculate the preparation time for two typical knitted products,namely,Polo-Shirt and T-Shirt with the case studies at General Textile Garment Joint Stock Company Hanoi and Star Fashion Company Limited.
文摘Software process is a framework for effective and timely delivery of software system. The framework plays a crucial role for software success. However, the development of large-scale software still faces the crisis of high risks, low quality, high costs and long cycle time. This paper proposed a three-phase parallel-pipelining software process model for improving speed and productivity, and reducing software costs and risks without sacrificing software quality. In this model, two strategies were presented. One strategy, based on subsystem-cost priority, was used to prevent software development cost wasting and to reduce software complexity as well; the other strategy, used for balancing subsystem complexity, was designed to reduce the software complexity in the later development stages. Moreover, the proposed function-detailed and workload-simplified subsystem pipelining software process model presents much higher parallelity than the concurrent incremental model. Finally, the component-based product line technology not only ensures software quality and further reduces cycle time, software costs, and software risks but also sufficiently and rationally utilizes previous software product resources and enhances the competition ability of software development organizations.
文摘The Goel-Okumoto software reliability model is one of the earliest attempts to use a non-homogeneous Poisson process to model failure times observed during software test interval. The model is known as exponential NHPP model as it describes exponential software failure curve. Parameter estimation, model fit and predictive analyses based on one sample have been conducted on the Goel-Okumoto software reliability model. However, predictive analyses based on two samples have not been conducted on the model. In two-sample prediction, the parameters and characteristics of the first sample are used to analyze and to make predictions for the second sample. This helps in saving time and resources during the software development process. This paper presents some results about predictive analyses for the Goel-Okumoto software reliability model based on two samples. We have addressed three issues in two-sample prediction associated closely with software development testing process. Bayesian methods based on non-informative priors have been adopted to develop solutions to these issues. The developed methodologies have been illustrated by two sets of software failure data simulated from the Goel-Okumoto software reliability model.