ISSN 1842-4562
Member of DOAJ
Volume 17 - 2022
Volume 16 - 2021
Volume 15 - 2020
Volume 14 - 2019
Volume 13 - 2018
Volume 12 - 2017
Volume 11 - 2016
Volume 10 - 2015
Volume 9 - 2014
Volume 8 - 2013
Volume 7 - 2012
Volume 6 - 2011
Volume 5 - 2010
Volume 4 - 2009
Volume 3 - 2008
Volume 2 - 2007
Volume 1 - 2006

Journal Home > Volume 2, Issue 4 - December 30, 2007

JAQM Volume 2, Issue 4 - December 30, 2007

Learning Organization and Team Learning. Systemic Approaching. Advantages and Challenges.


Structural Identification and Comparison of Intelligent Mobile Learning Environment (p 363)
Nitin UPADHYAY, Vishnu Prakash AGARWAL

This paper proposes a methodology using graph theory, matrix algebra and permanent function to compare different architecture (structure) design of intelligent mobile learning environment. The current work deals with the development/selection of optimum architecture (structural) model of iMLE. This can be done using the criterion as discussed in the paper.

The Learning Paradox and the University (p 375)
Constantin BRATIANU

Universities are by their nature learning based organizations. They deliver knowledge to the students through teaching processes. Students acquire knowledge through learning processes, from their professors and from other different knowledge resources. Since learning is a fundamental process within any university, people may consider universities as being learning organizations. This is a major error, especially in the former socialist countries. The purpose of this paper is to demonstrate that the most universities are far away from being learning organizations, due to some organizational learning barriers.

Information Technology and the Cognitive Domain (p 387)
Gabriel ZAMFIR

The purpose of this article is to map the context within which learning could occur, that is, the organizational learning processes and structures that can create or improve learning in a learning organization. Such an approach produces definition for learning organization and integrates the basis concepts into a model of organizational learning in the technologically environment, based on assisted instruction.

Comparing Organizational Learning Rates in Public and Non-Profit Schools in Qom Province of Iran (p 396)
Hassan ZAREI MATIN, Gholamreza JANDAGHI, Boshra MOINI

Regarding the increased complexity and dynamics of environmental factors and rapid changes, traditional organizations are not longer able to match with such changes and are destroying. Hence, as a tool for survival and matching with these changes, learning organizations are highly considered by many firms and corporations. What you are reading, is a summary of theoretical basics and findings of a research about the rate of organizational learning of schools. Present research is based on reviewing the rate of organizational learning of public and non-profit high schools. To obtain the research aims, this hypothesis was considered: “there is a difference between the rate of organizational learning of public and non-profit schools.” To collect the data and find the characteristics of learning organization (high school) library method and to review the rate of organizational learning, a structured questionnaire (after determining the reliability and validity) were utilized. Collected data were analyzed by using descriptive statistical method including frequency indices, frequency percent, mean and inferencial statistics such as Mann-Whitney test. Findings of Mann-Whitney test show that there is a meaningful difference between the rate of organizational learning of public and non-profit schools: “in all achieved characteristics, the rate of organizational learning of non-profit schools is more than public schools.”

Assessment of Organizational Learning within Teams (p 409)

Two representative approaches to measuring and assessing organizational learning are compared. Based on the benefits and drawbacks of each, an alternative framework is proposed for an assessment based on action science, from pragmatic point of view of members of a team within an organization which has declared an intention to improve learning processes.

Models and Systems for Structurization of Knowledge in Training (p 418)
Nicolae PELIN, Serghei PELIN

In this work the problems of the automated structurization and activation of the knowledge, saved and used by mankind, during the organization and training, and also that knowledge which are generated by experts (including teachers) in the current activity, are analyzed. The purpose - the further perfection of methods and systems of the automated structurization of knowledge and their applications for creation: the automated workplace of the lecturer (environment for the automated structurization of a lecture material), the virtual lecturer assistant (system for the automated dialogue with students), environments for preparation by students of reports, explanatory notes to course and degree works, the virtual adviser and knowledge tester for students testing during training, for other applications. Developed by authors "the matrix of elements of knowledge" - became acting system of the automated structurization of knowledge in training, it present interest for its use in other spheres.

Study of Equilibrium Prices, Using General Equilibrium Models (p 431)

The first part of the paper presents some theoretical aspects related to general equilibrium models and the motivation of formulating the general equilibrium models as mixed complementary ones. Then we present a general equilibrium application using GAMS software (General Algebraic Modelling Systems). The application is a general equilibrium model for the Romanian energetic system, considered as a component of the national economy.

The Effects of Non-Normality on Type Iii Error for Comparing Independent Means (p 444)

The major objective of this study was to investigate the effects of non-normality on Type III error rates for ANOVA F its three commonly recommended parametric counterparts namely Welch, Brown-Forsythe, and Alexander-Govern test. Therefore these tests were compared in terms of Type III error rates across the variety of population distributions, mean difference (effect size), and sample sizes. At the end of 100,000 simulation trials it was observed that the Type III error rates for four tests were affected by the effect size and sample size, whereas Type III errors were not affected from distribution shapes. Results of the simulation also indicated that increases in sample size and population mean difference decreased Type III error, and increased statistical test power. Across the all distributions, sample sizes and population mean differences, the Alexander-Govern test obtained higher estimates for power, lower estimates of Type III error (γ).

Automated Software Translation – Theoretical Background and Case Study (p 455)

The necessity for software migration is presented. The concept of software migrator is introduced. A generality metric for software translations is proposed. The business feasibility of automated versus manual migration is studied. nTile PHPtoJava, a working software migrator, is reviewed.

Harmonization Models for Designing Compound Systems (p 468)

A system to be designed and developed is composed of several sub-systems with complex configuration. The relationship between the sub-systems and the system cannot be fully expressed in analytical terms and has a high degree of uncertainty. Each sub-system can be designed and developed independently and is a subject of several possible measurable versions including both the cost of designing and creating the sub-system and its reliability. The problem is to assign reliability and cost requirements in the system design phase to all sub-systems, in order to: • achieve a specified reliability goal for the system, and • minimize the total costs of designing and creating of all the sub-systems. The corresponding dual problem is being solved as well. The third problem centers on optimizing the system’s structure in order to maximize the system’s utility by means of implementing local parametrical reliability and cost values.

Parallel Algorithms for Large Scale Macroeconometric Models (p 483)

Macroeconometric models with forward-looking variables give raise to very large systems of equations that requires heavy computations. These models was influenced by the development of new and efficient computational techniques and they are an interesting testing ground for the numerical methods addressed in this research. The most difficult problem in solving such models is to obtain the solution of the linear system that arises during the Newton step. For this purpose we have used both direct methods based on matrix factorization and nonstationary iterative methods, also called Krylov methods that provide an interesting alternative to the direct methods. In this paper we present performance results of both serial and parallel versions of the algorithms involved in solving these models. Although parallel implementation of the most dense linear algebra operations is a well understood process, the availability of general purpose, high performance parallel dense linear algebra libraries is limited by the complexity of implementation. This paper describes PLSS – (Parallel Linear System Solver) - a library which provides routines for linear system solving with an interface easy to use, that mirrors the natural description of sequential linear algebra algorithms.

Tendencies in the University System in Romania (p 494)
Tudorel ANDREI, Stelian STANCU, Irina Maria ISAIC-MANIU

Here is a short presentation of the main statistical indicators for higher education. The analysis was structured on national and regional level and also taking in account the differences between private and state universities. We have focused on some indicators regarding students and teachers flows. Romanian case is a special one, that had no private education during communist era and an exponential development after 1990.

Mathematical Model for Optimizing the Profit of the Pork Meat Chain (p 507)
Nicolae ISTUDOR, Victor MANOLE, Raluca Andreea ION, Toader PIRJOL

The research of the present study answer the question whether using measures for reducing the losses on the pork meat chain enables acquiring a higher level of effectiveness. In pursuing this question, a mathematical model for optimizing the profit of the chain is elaborated and implemented in two situations: scenario 1 – without measures for preventing the losses, and scenario 2 – with measures for preventing the losses. The results show that in the second situation the level of profit is higher, because of reducing the losses by applying measures for their prevention: farms’ re-technology, implementing an automatically system for feeding the animals, ensuring medication through feeding system, improving ventilation system, using devices for monitoring continuously the microclimate, establishing the optimal supply using scientifically methods etc. The conclusions have strong implications for chain operators who may acquire a higher level of profit by applying these measures of reducing the losses.

The Incidence of Pectoris Angina Post-Acute Myocardial Infarction in Romania – Ro-Stemi Database (p 517)
Catalina Liliana CALIN, Crina SINESCU, Claudiu HERTELIU

Cardio-vascular diseases are the major cause of mortality, the first role had ischemic cardiac disease follow-up by stroke. Each of these diseases account for some 15 million deaths per year that is 30% from the total number of death. The highest mortality is registered in the countries of former USSR and those of east Europe. France and Japan had the lowest rate of mortality by ischemic heart disease. This study is based on a retrospective analyse of the medical informationfrom provided by RO-STEMI database.

Data Mining Techniques in Processing Medical Knowledge (p 524)
Luminita STATE, Catalina COCIANU

Data mining is an evolving and growing area of research and development, both in academia as well as in industry. It involves interdisciplinary research and development encompassing diverse domains. In this age of multimedia data exploration, data mining should no longer be restricted to the mining of knowledge from large volumes of high-dimensional data sets in traditional databases only. The aim of the paper is to develop a new learning by examples PCA-based algorithm for extracting skeleton information from data to assure both good recognition performances, and generalization capabilities in case of large data set. The classes are represented in the measurement/feature space by continuous repartitions, that is the model is given by the family of density functions , where H stands for the finite set of hypothesis (classes). The basis of the learning process is represented by samples of possible different sizes coming from the considered classes. The skeleton of each class is given by the principal components obtained for the corresponding sample.

Book Review (p 533)

Book Review on
Published in “Studies in Fuzziness and Soft Computing”,
Springer, 2007