Technology in conservation: towards a system for in-field drone detection of invasive vegetation
- James, Katherine Margaret Frances
- Authors: James, Katherine Margaret Frances
- Date: 2020
- Subjects: Drone aircraft in remote sensing , Neural networks (Computer science) , Drone aircraft in remote sensing -- Case studies , Machine learning , Computer vision , Environmental monitoring -- Remote sensing , Invasive plants -- Monitoring
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/143408 , vital:38244
- Description: Remote sensing can assist in monitoring the spread of invasive vegetation. The adoption of camera-carrying unmanned aerial vehicles, commonly referred to as drones, as remote sensing tools has yielded images of higher spatial resolution than traditional techniques. Drones also have the potential to interact with the environment through the delivery of bio-control or herbicide, as seen with their adoption in precision agriculture. Unlike in agricultural applications, however, invasive plants do not have a predictable position relative to each other within the environment. To facilitate the adoption of drones as an environmental monitoring and management tool, drones need to be able to intelligently distinguish between invasive and non-invasive vegetation on the fly. In this thesis, we present the augmentation of a commercially available drone with a deep machine learning model to investigate the viability of differentiating between an invasive shrub and other vegetation. As a case study, this was applied to the shrub genus Hakea, originating in Australia and invasive in several countries including South Africa. However, for this research, the methodology is important, rather than the chosen target plant. A dataset was collected using the available drone and manually annotated to facilitate the supervised training of the model. Two approaches were explored, namely, classification and semantic segmentation. For each of these, several models were trained and evaluated to find the optimal one. The chosen model was then interfaced with the drone via an Android application on a mobile device and its performance was preliminarily evaluated in the field. Based on these findings, refinements were made and thereafter a thorough field evaluation was performed to determine the best conditions for model operation. Results from the classification task show that deep learning models are capable of distinguishing between target and other shrubs in ideal candidate windows. However, classification in this manner is restricted by the proposal of such candidate windows. End-to-end image segmentation using deep learning overcomes this problem, classifying the image in a pixel-wise manner. Furthermore, the use of appropriate loss functions was found to improve model performance. Field tests show that illumination and shadow pose challenges to the model, but that good recall can be achieved when the conditions are ideal. False positive detection remains an issue that could be improved. This approach shows the potential for drones as an environmental monitoring and management tool when coupled with deep machine learning techniques and outlines potential problems that may be encountered.
- Full Text:
- Date Issued: 2020
- Authors: James, Katherine Margaret Frances
- Date: 2020
- Subjects: Drone aircraft in remote sensing , Neural networks (Computer science) , Drone aircraft in remote sensing -- Case studies , Machine learning , Computer vision , Environmental monitoring -- Remote sensing , Invasive plants -- Monitoring
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/143408 , vital:38244
- Description: Remote sensing can assist in monitoring the spread of invasive vegetation. The adoption of camera-carrying unmanned aerial vehicles, commonly referred to as drones, as remote sensing tools has yielded images of higher spatial resolution than traditional techniques. Drones also have the potential to interact with the environment through the delivery of bio-control or herbicide, as seen with their adoption in precision agriculture. Unlike in agricultural applications, however, invasive plants do not have a predictable position relative to each other within the environment. To facilitate the adoption of drones as an environmental monitoring and management tool, drones need to be able to intelligently distinguish between invasive and non-invasive vegetation on the fly. In this thesis, we present the augmentation of a commercially available drone with a deep machine learning model to investigate the viability of differentiating between an invasive shrub and other vegetation. As a case study, this was applied to the shrub genus Hakea, originating in Australia and invasive in several countries including South Africa. However, for this research, the methodology is important, rather than the chosen target plant. A dataset was collected using the available drone and manually annotated to facilitate the supervised training of the model. Two approaches were explored, namely, classification and semantic segmentation. For each of these, several models were trained and evaluated to find the optimal one. The chosen model was then interfaced with the drone via an Android application on a mobile device and its performance was preliminarily evaluated in the field. Based on these findings, refinements were made and thereafter a thorough field evaluation was performed to determine the best conditions for model operation. Results from the classification task show that deep learning models are capable of distinguishing between target and other shrubs in ideal candidate windows. However, classification in this manner is restricted by the proposal of such candidate windows. End-to-end image segmentation using deep learning overcomes this problem, classifying the image in a pixel-wise manner. Furthermore, the use of appropriate loss functions was found to improve model performance. Field tests show that illumination and shadow pose challenges to the model, but that good recall can be achieved when the conditions are ideal. False positive detection remains an issue that could be improved. This approach shows the potential for drones as an environmental monitoring and management tool when coupled with deep machine learning techniques and outlines potential problems that may be encountered.
- Full Text:
- Date Issued: 2020
The analysis of a computer music network and the implementation of essential subsystems
- Authors: Wilks, Antony John
- Date: 1995
- Subjects: Computer networks , Computer music , MIDI (Standard)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4666 , http://hdl.handle.net/10962/d1006682 , Computer networks , Computer music , MIDI (Standard)
- Description: The inability to share resources in commercial and institutional computer music studios results in non-optimal resource utilisation. The use of computers to process, store and communicate data can be extended within these studios, to provide the capability of sharing resources amongst their users. This thesis describes a computer music network which was designed for this purpose. Certain devices had to be custom built for the implementation of the network. The thesis discusses the design and construction of these devices.
- Full Text:
- Date Issued: 1995
- Authors: Wilks, Antony John
- Date: 1995
- Subjects: Computer networks , Computer music , MIDI (Standard)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4666 , http://hdl.handle.net/10962/d1006682 , Computer networks , Computer music , MIDI (Standard)
- Description: The inability to share resources in commercial and institutional computer music studios results in non-optimal resource utilisation. The use of computers to process, store and communicate data can be extended within these studios, to provide the capability of sharing resources amongst their users. This thesis describes a computer music network which was designed for this purpose. Certain devices had to be custom built for the implementation of the network. The thesis discusses the design and construction of these devices.
- Full Text:
- Date Issued: 1995
The automatic generation of code generators with particular reference to cobol
- Authors: Bulmer, Allan Roy
- Date: 1980 , 2013-03-15
- Subjects: Code generators , COBOL (Computer program language)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4613 , http://hdl.handle.net/10962/d1004913 , Code generators , COBOL (Computer program language)
- Full Text:
- Date Issued: 1980
- Authors: Bulmer, Allan Roy
- Date: 1980 , 2013-03-15
- Subjects: Code generators , COBOL (Computer program language)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4613 , http://hdl.handle.net/10962/d1004913 , Code generators , COBOL (Computer program language)
- Full Text:
- Date Issued: 1980
The design and implementation of a fourth generation programming language
- Authors: Iverson, Carn Martin
- Date: 1989 , 2013-03-04
- Subjects: Programming languages (Electronic computers) , Fourth generation computers
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4584 , http://hdl.handle.net/10962/d1004652 , Programming languages (Electronic computers) , Fourth generation computers
- Description: IV is a very high level language designed for use in a real time production control environment. While most fourth generation languages are intended for use by end users, IV is more suitable for skilled professional programmers. One of the major design objectives of IV is a dramatic improvement in programmer efficiency during application program development. Non-procedural constructs provided by the language and the use of a number of interactive development tools provide an environment for achieving this goal. This report presents a language proposal for IV, and addresses related design and implementation issues.
- Full Text:
- Date Issued: 1989
- Authors: Iverson, Carn Martin
- Date: 1989 , 2013-03-04
- Subjects: Programming languages (Electronic computers) , Fourth generation computers
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4584 , http://hdl.handle.net/10962/d1004652 , Programming languages (Electronic computers) , Fourth generation computers
- Description: IV is a very high level language designed for use in a real time production control environment. While most fourth generation languages are intended for use by end users, IV is more suitable for skilled professional programmers. One of the major design objectives of IV is a dramatic improvement in programmer efficiency during application program development. Non-procedural constructs provided by the language and the use of a number of interactive development tools provide an environment for achieving this goal. This report presents a language proposal for IV, and addresses related design and implementation issues.
- Full Text:
- Date Issued: 1989
The design, development and evaluation of cross-platform mobile applications and services supporting social accountability monitoring
- Authors: Reynell, Edward Robin
- Date: 2016
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/3652 , vital:20533
- Description: Local government processes require meaningful and effective participation from both citizens and their governments in order to remain truly democratic. This project investigates the use of mobile phones as a tool for supporting this participation. MobiSAM, a system which aims to enhance the Social Accountability Monitoring (SAM) methodology at local government level, has been designed and implemented. The research presented in this thesis examines tools and techniques for the development of cross-platform client applications, allowing access to the MobiSAM service, across heterogeneous mobile platforms, handsets and interaction styles. Particular attention is paid to providing an easily navigated user interface (UI), as well as offering clear and concise visualisation capabilities. Depending on the host device, interactivity is also included within these visualisations, potentially helping provide further insight into the visualised data. Guided by the results obtained from a comprehensive baseline study of the Grahamstown area, steps are taken in an attempt to lower the barrier of entry to using the MobiSAM service, potentially maximising its market reach. These include extending client application support to all identified mobile platforms (including feature phones); providing multi-language UIs (in English, isiXhosa and Afrikaans); as well as ensuring client application data usage is kept to a minimum. The particular strengths of a given device are also leveraged, such as its camera capabilities and built-in Global Positioning System (GPS) module, potentially allowing for more effective engagement with local municipalities. Additionally, a Short Message Service (SMS) gateway is developed, allowing all Global System for Mobile Communications (GSM) compatible handsets access to the MobiSAM service via traditional SMS. Following an iterative, user-centred design process, a thorough evaluation of the client application is also performed, in an attempt to gather feedback relating to the navigation and visualisation capabilities. The results of which are used to further refine its design. A comparative usability evaluation using two different versions of the cross-platform client application is also undertaken, highlighting the perceived memorability, learnabilitv and satisfaction of each. Results from the evaluation reveals which version of the client application is to be deployed during future pilot studies.
- Full Text:
- Date Issued: 2016
- Authors: Reynell, Edward Robin
- Date: 2016
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/3652 , vital:20533
- Description: Local government processes require meaningful and effective participation from both citizens and their governments in order to remain truly democratic. This project investigates the use of mobile phones as a tool for supporting this participation. MobiSAM, a system which aims to enhance the Social Accountability Monitoring (SAM) methodology at local government level, has been designed and implemented. The research presented in this thesis examines tools and techniques for the development of cross-platform client applications, allowing access to the MobiSAM service, across heterogeneous mobile platforms, handsets and interaction styles. Particular attention is paid to providing an easily navigated user interface (UI), as well as offering clear and concise visualisation capabilities. Depending on the host device, interactivity is also included within these visualisations, potentially helping provide further insight into the visualised data. Guided by the results obtained from a comprehensive baseline study of the Grahamstown area, steps are taken in an attempt to lower the barrier of entry to using the MobiSAM service, potentially maximising its market reach. These include extending client application support to all identified mobile platforms (including feature phones); providing multi-language UIs (in English, isiXhosa and Afrikaans); as well as ensuring client application data usage is kept to a minimum. The particular strengths of a given device are also leveraged, such as its camera capabilities and built-in Global Positioning System (GPS) module, potentially allowing for more effective engagement with local municipalities. Additionally, a Short Message Service (SMS) gateway is developed, allowing all Global System for Mobile Communications (GSM) compatible handsets access to the MobiSAM service via traditional SMS. Following an iterative, user-centred design process, a thorough evaluation of the client application is also performed, in an attempt to gather feedback relating to the navigation and visualisation capabilities. The results of which are used to further refine its design. A comparative usability evaluation using two different versions of the cross-platform client application is also undertaken, highlighting the perceived memorability, learnabilitv and satisfaction of each. Results from the evaluation reveals which version of the client application is to be deployed during future pilot studies.
- Full Text:
- Date Issued: 2016
The development and evaluation of a custom-built synchronous online learning environment for tertiary education in South Africa
- Authors: Halse, Michelle Louise
- Date: 2008 , 2008-02-23
- Subjects: Education, Higher -- South Africa -- Computer-assisted instruction , Universities and colleges -- Computer networks -- South Africa , Internet in higher education -- South Africa , Distance education -- South Africa -- Computer-assisted instruction , Computer-assisted instruction -- South Africa , Computer science -- Study and teaching (Higher) -- South Africa
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4636 , http://hdl.handle.net/10962/d1006545 , Education, Higher -- South Africa -- Computer-assisted instruction , Universities and colleges -- Computer networks -- South Africa , Internet in higher education -- South Africa , Distance education -- South Africa -- Computer-assisted instruction , Computer-assisted instruction -- South Africa , Computer science -- Study and teaching (Higher) -- South Africa
- Description: The Departments of Computer Science and Information Systems at Rhodes University currently share certain honours-level (fourth year) course modules with students from the corresponding departments at the previously disadvantaged University of Fort Hare. These lectures are currently delivered using video-conferencing. This was found to present a number of problems including challenges in terms of implementing desired pedagogical approaches, inequitable learning experiences, student disengagement at the remote venue, and inflexibility of the video-conferencing system. In order to address these problems, various e-learning modes were investigated and synchronous e-learning were found to offer a number of advantages over asynchronous e-learning. Live Virtual Classrooms (LVCs) were identified as synchronous e-learning tools that support the pedagogical principles important to the two universities and to the broader context of South African tertiary education, and commercial LVC applications were investigated and evaluated. Informed by the results of this investigation a small, simple LVC was designed, developed and customised for use in a predominantly academic sphere and deployment in a South African tertiary educational context. Testing and evaluation of this solution was carried out and the results analysed in terms of the LVC’s technical merits and the pedagogical value of the solution as experienced by students and lecturers/facilitators. An evaluation of this solution indicated that the LVC solves a number of the identified problems with video-conferencing and also provides a flexible/customisable/extensible solution that supports highly interactive, collaborative, learner-centred education. The custom LVC solution could be easily adapted to the specific needs of any tertiary educational institute in the country, and results may benefit other tertiary educational institutions involved in or dependant on distance learning.
- Full Text:
- Date Issued: 2008
- Authors: Halse, Michelle Louise
- Date: 2008 , 2008-02-23
- Subjects: Education, Higher -- South Africa -- Computer-assisted instruction , Universities and colleges -- Computer networks -- South Africa , Internet in higher education -- South Africa , Distance education -- South Africa -- Computer-assisted instruction , Computer-assisted instruction -- South Africa , Computer science -- Study and teaching (Higher) -- South Africa
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4636 , http://hdl.handle.net/10962/d1006545 , Education, Higher -- South Africa -- Computer-assisted instruction , Universities and colleges -- Computer networks -- South Africa , Internet in higher education -- South Africa , Distance education -- South Africa -- Computer-assisted instruction , Computer-assisted instruction -- South Africa , Computer science -- Study and teaching (Higher) -- South Africa
- Description: The Departments of Computer Science and Information Systems at Rhodes University currently share certain honours-level (fourth year) course modules with students from the corresponding departments at the previously disadvantaged University of Fort Hare. These lectures are currently delivered using video-conferencing. This was found to present a number of problems including challenges in terms of implementing desired pedagogical approaches, inequitable learning experiences, student disengagement at the remote venue, and inflexibility of the video-conferencing system. In order to address these problems, various e-learning modes were investigated and synchronous e-learning were found to offer a number of advantages over asynchronous e-learning. Live Virtual Classrooms (LVCs) were identified as synchronous e-learning tools that support the pedagogical principles important to the two universities and to the broader context of South African tertiary education, and commercial LVC applications were investigated and evaluated. Informed by the results of this investigation a small, simple LVC was designed, developed and customised for use in a predominantly academic sphere and deployment in a South African tertiary educational context. Testing and evaluation of this solution was carried out and the results analysed in terms of the LVC’s technical merits and the pedagogical value of the solution as experienced by students and lecturers/facilitators. An evaluation of this solution indicated that the LVC solves a number of the identified problems with video-conferencing and also provides a flexible/customisable/extensible solution that supports highly interactive, collaborative, learner-centred education. The custom LVC solution could be easily adapted to the specific needs of any tertiary educational institute in the country, and results may benefit other tertiary educational institutions involved in or dependant on distance learning.
- Full Text:
- Date Issued: 2008
The development of a discovery and control environment for networked audio devices based on a study of current audio control protocols
- Authors: Eales, Andrew Arnold
- Date: 2016
- Language: English
- Type: Thesis , Doctoral , PhD
- Identifier: http://hdl.handle.net/10962/539 , vital:19968
- Description: This dissertation develops a standard device model for networked audio devices and introduces a novel discovery and control environment that uses the developed device model. The proposed standard device model is derived from a study of current audio control protocols. Both the functional capabilities and design principles of audio control protocols are investigated with an emphasis on Open Sound Control, SNMP and IEC-62379, AES64, CopperLan and UPnP. An abstract model of networked audio devices is developed, and the model is implemented in each of the previously mentioned control protocols. This model is also used within a novel discovery and control environment designed around a distributed associative memory termed an object space. This environment challenges the accepted notions of the functionality provided by a control protocol. The study concludes by comparing the salient features of the different control protocols encountered in this study. Different approaches to control protocol design are considered, and several design heuristics for control protocols are proposed.
- Full Text:
- Date Issued: 2016
- Authors: Eales, Andrew Arnold
- Date: 2016
- Language: English
- Type: Thesis , Doctoral , PhD
- Identifier: http://hdl.handle.net/10962/539 , vital:19968
- Description: This dissertation develops a standard device model for networked audio devices and introduces a novel discovery and control environment that uses the developed device model. The proposed standard device model is derived from a study of current audio control protocols. Both the functional capabilities and design principles of audio control protocols are investigated with an emphasis on Open Sound Control, SNMP and IEC-62379, AES64, CopperLan and UPnP. An abstract model of networked audio devices is developed, and the model is implemented in each of the previously mentioned control protocols. This model is also used within a novel discovery and control environment designed around a distributed associative memory termed an object space. This environment challenges the accepted notions of the functionality provided by a control protocol. The study concludes by comparing the salient features of the different control protocols encountered in this study. Different approaches to control protocol design are considered, and several design heuristics for control protocols are proposed.
- Full Text:
- Date Issued: 2016
The implementation of a mobile application to decrease occupational sitting through goal setting and social comparison
- Authors: Tsaoane, Moipone Lipalesa
- Date: 2022-10-14
- Subjects: Uncatalogued
- Language: English
- Type: Academic theses , Master's theses , text
- Identifier: http://hdl.handle.net/10962/365544 , vital:65758
- Description: Background: Feedback proves to be a valuable tool in behaviour change as it is said to increase compliance and improve the effectiveness of interventions. Interventions that focus on decreasing sedentary behaviour as an independent factor from physical activity are necessary, especially for office workers who spend most of their day seated. There is insufficient knowledge regarding the effectiveness of feedback as a tool to decrease sedentary behaviour. This project implemented a tool that can be used to determine this. To take advantage of the cost-effectiveness and scalability of digital technologies, a mobile application was selected as the mode of delivery. Method: The application was designed as an intervention, using the Theoretical Domains Framework. It was then implemented into a fully functioning application through an agile development process, using Xam- arin.Forms framework. Due to challenges with this framework, a second application was developed using the React Native framework. Pilot studies were used for testing, with the final one consisting of Rhodes University employees. Results: The Xamarin.Forms application proved to be unfeasible; some users experienced fatal errors and crashes. The React Native application worked as desired and produced accurate and consistent step count readings, proving feasible from a functionality standpoint. The agile methodology enabled the developer to focus on implementing and testing one component at a time, which made the development process more manageable. Conclusion: Future work must conduct empirical studies to determine if feedback is an effective tool compared to a control group and which type of feedback (between goal-setting and social comparison) is most effective. , Thesis (MSc) -- Faculty of Science, Computer Science, 2022
- Full Text:
- Date Issued: 2022-10-14
- Authors: Tsaoane, Moipone Lipalesa
- Date: 2022-10-14
- Subjects: Uncatalogued
- Language: English
- Type: Academic theses , Master's theses , text
- Identifier: http://hdl.handle.net/10962/365544 , vital:65758
- Description: Background: Feedback proves to be a valuable tool in behaviour change as it is said to increase compliance and improve the effectiveness of interventions. Interventions that focus on decreasing sedentary behaviour as an independent factor from physical activity are necessary, especially for office workers who spend most of their day seated. There is insufficient knowledge regarding the effectiveness of feedback as a tool to decrease sedentary behaviour. This project implemented a tool that can be used to determine this. To take advantage of the cost-effectiveness and scalability of digital technologies, a mobile application was selected as the mode of delivery. Method: The application was designed as an intervention, using the Theoretical Domains Framework. It was then implemented into a fully functioning application through an agile development process, using Xam- arin.Forms framework. Due to challenges with this framework, a second application was developed using the React Native framework. Pilot studies were used for testing, with the final one consisting of Rhodes University employees. Results: The Xamarin.Forms application proved to be unfeasible; some users experienced fatal errors and crashes. The React Native application worked as desired and produced accurate and consistent step count readings, proving feasible from a functionality standpoint. The agile methodology enabled the developer to focus on implementing and testing one component at a time, which made the development process more manageable. Conclusion: Future work must conduct empirical studies to determine if feedback is an effective tool compared to a control group and which type of feedback (between goal-setting and social comparison) is most effective. , Thesis (MSc) -- Faculty of Science, Computer Science, 2022
- Full Text:
- Date Issued: 2022-10-14
The monitor and synchroniser concepts in the programming language CLANG
- Authors: Chalmers, Alan Gordon
- Date: 1985
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4616 , http://hdl.handle.net/10962/d1006132
- Full Text:
- Date Issued: 1985
- Authors: Chalmers, Alan Gordon
- Date: 1985
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4616 , http://hdl.handle.net/10962/d1006132
- Full Text:
- Date Issued: 1985
The P.R.O. expert system shell
- Authors: Bradshaw, John
- Date: 1987 , 2013-04-03
- Subjects: Expert systems (Computer science)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4617 , http://hdl.handle.net/10962/d1006302 , Expert systems (Computer science)
- Description: This thesis reports the research which led to the development of the P.R .O. Expert System Shell. The P.R.O . System is primarily, though not exclusively , designed for use in ecological domains. In the light of two specific expert systems, The RCS (River Conservation Status) and the Aquaculture Systems, which were developed as part of this research, a number of areas of importance have been identified. The most significant of these is the need to handle uncertainty effectively. The style of knowledge representation to be implemented also plays an important role. After consulting the relevant literature and the available microcomputer expert system shells, a number of ideas have been included in the P.R.O. System. The P.R.O . System is a backward chaining, production system based expert system shell. It embodies a simple but effective method of handling uncertainty. An important feature of this method is that it takes cognizance of the different relative importances of the conditions which need to be satisfied before a conclusion can be reached. The knowledge base consists of more than rules and questions. It also contains meta-knowledge, which is used by the inference engine. The P.R.O. System has been designed to be of practical use. Its strongest recommendations are therefore, that the two non-trivial systems which have been implemented in it, have been accepted by the experts and their peers as systems which produce good, accurate answers . , KMBT_363 , Adobe Acrobat 9.53 Paper Capture Plug-in
- Full Text:
- Date Issued: 1987
- Authors: Bradshaw, John
- Date: 1987 , 2013-04-03
- Subjects: Expert systems (Computer science)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4617 , http://hdl.handle.net/10962/d1006302 , Expert systems (Computer science)
- Description: This thesis reports the research which led to the development of the P.R .O. Expert System Shell. The P.R.O . System is primarily, though not exclusively , designed for use in ecological domains. In the light of two specific expert systems, The RCS (River Conservation Status) and the Aquaculture Systems, which were developed as part of this research, a number of areas of importance have been identified. The most significant of these is the need to handle uncertainty effectively. The style of knowledge representation to be implemented also plays an important role. After consulting the relevant literature and the available microcomputer expert system shells, a number of ideas have been included in the P.R.O. System. The P.R.O . System is a backward chaining, production system based expert system shell. It embodies a simple but effective method of handling uncertainty. An important feature of this method is that it takes cognizance of the different relative importances of the conditions which need to be satisfied before a conclusion can be reached. The knowledge base consists of more than rules and questions. It also contains meta-knowledge, which is used by the inference engine. The P.R.O. System has been designed to be of practical use. Its strongest recommendations are therefore, that the two non-trivial systems which have been implemented in it, have been accepted by the experts and their peers as systems which produce good, accurate answers . , KMBT_363 , Adobe Acrobat 9.53 Paper Capture Plug-in
- Full Text:
- Date Issued: 1987
The remote configuration of devices within home entertainment networks
- Authors: Dembovsky, Colin
- Date: 2002
- Subjects: Home entertainment systems , Home video systems
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4685 , http://hdl.handle.net/10962/d1007795 , Home entertainment systems , Home video systems
- Description: This thesis examines home entertainment network remote configuration solutions. It does so by inspecting four home entertainment networking solution specifications - HAVi, Jini, AV/C and UPnP. Two of these (AV/C and UPnP) are implemented partially for a system allowing a TV to configure an AudioNideo Receiver (AV/R) remotely on the network (a process known as remote configuration). The two implementations are then more closely investigated and several implementation differences in the approach between the remote configuration method of device configuration and other methods of device configuration are discerned. These different approaches are then categorised into one of two theoretical models of communication for configuring devices on home entertainment networks - the Rendering model and the Programmed model. By classifying a particular method of device configuration into one of the two models, manufacturers can quickly determine the inherent strengths and weaknesses of that method , KMBT_363 , Adobe Acrobat 9.54 Paper Capture Plug-in
- Full Text:
- Authors: Dembovsky, Colin
- Date: 2002
- Subjects: Home entertainment systems , Home video systems
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4685 , http://hdl.handle.net/10962/d1007795 , Home entertainment systems , Home video systems
- Description: This thesis examines home entertainment network remote configuration solutions. It does so by inspecting four home entertainment networking solution specifications - HAVi, Jini, AV/C and UPnP. Two of these (AV/C and UPnP) are implemented partially for a system allowing a TV to configure an AudioNideo Receiver (AV/R) remotely on the network (a process known as remote configuration). The two implementations are then more closely investigated and several implementation differences in the approach between the remote configuration method of device configuration and other methods of device configuration are discerned. These different approaches are then categorised into one of two theoretical models of communication for configuring devices on home entertainment networks - the Rendering model and the Programmed model. By classifying a particular method of device configuration into one of the two models, manufacturers can quickly determine the inherent strengths and weaknesses of that method , KMBT_363 , Adobe Acrobat 9.54 Paper Capture Plug-in
- Full Text:
The role of computational thinking in introductory computer science
- Authors: Gouws, Lindsey Ann
- Date: 2014
- Subjects: Computer science , Computational complexity , Problem solving -- Study and teaching
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4690 , http://hdl.handle.net/10962/d1011152 , Computer science , Computational complexity , Problem solving -- Study and teaching
- Description: Computational thinking (CT) is gaining recognition as an important skill for students, both in computer science and other disciplines. Although there has been much focus on this field in recent years, it is rarely taught as a formal course, and there is little consensus on what exactly CT entails and how to teach and evaluate it. This research addresses the lack of resources for integrating CT into the introductory computer science curriculum. The question that we aim to answer is whether CT can be evaluated in a meaningful way. A CT framework that outlines the skills and techniques comprising CT and describes the nature of student engagement was developed; this is used as the basis for this research. An assessment (CT test) was then created to gauge the ability of incoming students, and a CT-specfic computer game was developed based on the analysis of an existing game. A set of problem solving strategies and practice activities were then recommended based on criteria defined in the framework. The results revealed that the CT abilities of first year university students are relatively poor, but that the students' scores for the CT test could be used as a predictor for their future success in computer science courses. The framework developed for this research proved successful when applied to the test, computer game evaluation, and classification of strategies and activities. Through this research, we established that CT is a skill that first year computer science students are lacking, and that using CT exercises alongside traditional programming instruction can improve students' learning experiences.
- Full Text:
- Date Issued: 2014
- Authors: Gouws, Lindsey Ann
- Date: 2014
- Subjects: Computer science , Computational complexity , Problem solving -- Study and teaching
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4690 , http://hdl.handle.net/10962/d1011152 , Computer science , Computational complexity , Problem solving -- Study and teaching
- Description: Computational thinking (CT) is gaining recognition as an important skill for students, both in computer science and other disciplines. Although there has been much focus on this field in recent years, it is rarely taught as a formal course, and there is little consensus on what exactly CT entails and how to teach and evaluate it. This research addresses the lack of resources for integrating CT into the introductory computer science curriculum. The question that we aim to answer is whether CT can be evaluated in a meaningful way. A CT framework that outlines the skills and techniques comprising CT and describes the nature of student engagement was developed; this is used as the basis for this research. An assessment (CT test) was then created to gauge the ability of incoming students, and a CT-specfic computer game was developed based on the analysis of an existing game. A set of problem solving strategies and practice activities were then recommended based on criteria defined in the framework. The results revealed that the CT abilities of first year university students are relatively poor, but that the students' scores for the CT test could be used as a predictor for their future success in computer science courses. The framework developed for this research proved successful when applied to the test, computer game evaluation, and classification of strategies and activities. Through this research, we established that CT is a skill that first year computer science students are lacking, and that using CT exercises alongside traditional programming instruction can improve students' learning experiences.
- Full Text:
- Date Issued: 2014
The synthesis of sound with application in a MIDI environment
- Authors: Kesterton, Anthony James
- Date: 1991
- Subjects: Computer sound processing -- Research , Music -- Data processing -- Research , MIDI (Standard)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4678 , http://hdl.handle.net/10962/d1006701 , Computer sound processing -- Research , Music -- Data processing -- Research , MIDI (Standard)
- Description: The wide range of options for experimentation with the synthesis of sound are usually expensive, difficult to obtain, or limit the experimenter. The work described in this thesis shows how the IBM PC and software can be combined to provide a suitable platform for experimentation with different synthesis techniques. This platform is based on the PC, the Musical Instrument Digital Interface (MIDI) and a musical instrument called a digital sampler. The fundamental concepts of sound are described, with reference to digital sound reproduction. A number of synthesis techniques are described. These are evaluated according to the criteria of generality, efficiency and control. The techniques discussed are additive synthesis, frequency modulation synthesis, subtractive synthesis, granular synthesis, resynthesis, wavetable synthesis, and sampling. Spiral synthesis, physical modelling, waveshaping and spectral interpolation are discussed briefly. The Musical Instrument Digital Interface is a standard method of connecting digital musical instruments together. It is the MIDI standard and equipment conforming to that standard that makes this implementation of synthesis techniques possible. As a demonstration of the PC platform, additive synthesis, frequency modulation synthesis, granular synthesis and spiral synthesis have been implemented in software. A PC equipped with a MIDI interface card is used to perform the synthesis. The MIDI protocol is used to transmit the resultant sound to a digital sampler. The INMOS transputer is used as an accelerator, as the calculation of a waveform using software is a computational intensive process. It is concluded that sound synthesis can be performed successfully using a PC and the appropriate software, and utilizing the facilities provided by a MIDI environment including a digital sampler.
- Full Text:
- Date Issued: 1991
- Authors: Kesterton, Anthony James
- Date: 1991
- Subjects: Computer sound processing -- Research , Music -- Data processing -- Research , MIDI (Standard)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4678 , http://hdl.handle.net/10962/d1006701 , Computer sound processing -- Research , Music -- Data processing -- Research , MIDI (Standard)
- Description: The wide range of options for experimentation with the synthesis of sound are usually expensive, difficult to obtain, or limit the experimenter. The work described in this thesis shows how the IBM PC and software can be combined to provide a suitable platform for experimentation with different synthesis techniques. This platform is based on the PC, the Musical Instrument Digital Interface (MIDI) and a musical instrument called a digital sampler. The fundamental concepts of sound are described, with reference to digital sound reproduction. A number of synthesis techniques are described. These are evaluated according to the criteria of generality, efficiency and control. The techniques discussed are additive synthesis, frequency modulation synthesis, subtractive synthesis, granular synthesis, resynthesis, wavetable synthesis, and sampling. Spiral synthesis, physical modelling, waveshaping and spectral interpolation are discussed briefly. The Musical Instrument Digital Interface is a standard method of connecting digital musical instruments together. It is the MIDI standard and equipment conforming to that standard that makes this implementation of synthesis techniques possible. As a demonstration of the PC platform, additive synthesis, frequency modulation synthesis, granular synthesis and spiral synthesis have been implemented in software. A PC equipped with a MIDI interface card is used to perform the synthesis. The MIDI protocol is used to transmit the resultant sound to a digital sampler. The INMOS transputer is used as an accelerator, as the calculation of a waveform using software is a computational intensive process. It is concluded that sound synthesis can be performed successfully using a PC and the appropriate software, and utilizing the facilities provided by a MIDI environment including a digital sampler.
- Full Text:
- Date Issued: 1991
Toward an automated botnet analysis framework: a darkcomet case-study
- Authors: du Bruyn, Jeremy Cecil
- Date: 2016
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/2937 , vital:20344
- Full Text:
- Date Issued: 2016
- Authors: du Bruyn, Jeremy Cecil
- Date: 2016
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/2937 , vital:20344
- Full Text:
- Date Issued: 2016
Towards a capability maturity model for a cyber range
- Authors: Aschmann, Michael Joseph
- Date: 2020
- Subjects: Computer software -- Development , Computer security
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/163142 , vital:41013
- Description: This work describes research undertaken towards the development of a Capability Maturity Model (CMM) for Cyber Ranges (CRs) focused on cyber security. Global cyber security needs are on the rise, and the need for attribution within the cyber domain is of particular concern. This has prompted major efforts to enhance cyber capabilities within organisations to increase their total cyber resilience posture. These efforts include, but are not limited to, the testing of computational devices, networks, and applications, and cyber skills training focused on prevention, detection and cyber attack response. A cyber range allows for the testing of the computational environment. By developing cyber events within a confined virtual or sand-boxed cyber environment, a cyber range can prepare the next generation of cyber security specialists to handle a variety of potential cyber attacks. Cyber ranges have different purposes, each designed to fulfil a different computational testing and cyber training goal; consequently, cyber ranges can vary greatly in the level of variety, capability, maturity and complexity. As cyber ranges proliferate and become more and more valued as tools for cyber security, a method to classify or rate them becomes essential. Yet while a universal criteria for measuring cyber ranges in terms of their capability maturity levels becomes more critical, there are currently very limited resources for researchers aiming to perform this kind of work. For this reason, this work proposes and describes a CMM, designed to give organisations the ability to benchmark the capability maturity of a given cyber range. This research adopted a synthesised approach to the development of a CMM, grounded in prior research and focused on the production of a conceptual model that provides a useful level of abstraction. In order to achieve this goal, the core capability elements of a cyber range are defined with their relative importance, allowing for the development of a proposed classification cyber range levels. An analysis of data gathered during the course of an expert review, together with other research, further supported the development of the conceptual model. In the context of cyber range capability, classification will include the ability of the cyber range to perform its functions optimally with different core capability elements, focusing on the Measurement of Capability (MoC) with its elements, namely effect, performance and threat ability. Cyber range maturity can evolve over time and can be defined through the Measurement of Maturity (MoM) with its elements, namely people, processes, technology. The combination of these measurements utilising the CMM for a CR determines the capability maturity level of a CR. The primary outcome of this research is the proposed level-based CMM framework for a cyber range, developed using adopted and synthesised CMMs, the analysis of an expert review, and the mapping of the results.
- Full Text:
- Date Issued: 2020
- Authors: Aschmann, Michael Joseph
- Date: 2020
- Subjects: Computer software -- Development , Computer security
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/163142 , vital:41013
- Description: This work describes research undertaken towards the development of a Capability Maturity Model (CMM) for Cyber Ranges (CRs) focused on cyber security. Global cyber security needs are on the rise, and the need for attribution within the cyber domain is of particular concern. This has prompted major efforts to enhance cyber capabilities within organisations to increase their total cyber resilience posture. These efforts include, but are not limited to, the testing of computational devices, networks, and applications, and cyber skills training focused on prevention, detection and cyber attack response. A cyber range allows for the testing of the computational environment. By developing cyber events within a confined virtual or sand-boxed cyber environment, a cyber range can prepare the next generation of cyber security specialists to handle a variety of potential cyber attacks. Cyber ranges have different purposes, each designed to fulfil a different computational testing and cyber training goal; consequently, cyber ranges can vary greatly in the level of variety, capability, maturity and complexity. As cyber ranges proliferate and become more and more valued as tools for cyber security, a method to classify or rate them becomes essential. Yet while a universal criteria for measuring cyber ranges in terms of their capability maturity levels becomes more critical, there are currently very limited resources for researchers aiming to perform this kind of work. For this reason, this work proposes and describes a CMM, designed to give organisations the ability to benchmark the capability maturity of a given cyber range. This research adopted a synthesised approach to the development of a CMM, grounded in prior research and focused on the production of a conceptual model that provides a useful level of abstraction. In order to achieve this goal, the core capability elements of a cyber range are defined with their relative importance, allowing for the development of a proposed classification cyber range levels. An analysis of data gathered during the course of an expert review, together with other research, further supported the development of the conceptual model. In the context of cyber range capability, classification will include the ability of the cyber range to perform its functions optimally with different core capability elements, focusing on the Measurement of Capability (MoC) with its elements, namely effect, performance and threat ability. Cyber range maturity can evolve over time and can be defined through the Measurement of Maturity (MoM) with its elements, namely people, processes, technology. The combination of these measurements utilising the CMM for a CR determines the capability maturity level of a CR. The primary outcome of this research is the proposed level-based CMM framework for a cyber range, developed using adopted and synthesised CMMs, the analysis of an expert review, and the mapping of the results.
- Full Text:
- Date Issued: 2020
Towards a collection of cost-effective technologies in support of the NIST cybersecurity framework
- Shackleton, Bruce Michael Stuart
- Authors: Shackleton, Bruce Michael Stuart
- Date: 2018
- Subjects: National Institute of Standards and Technology (U.S.) , Computer security , Computer networks Security measures , Small business Information technology Cost effectiveness , Open source software
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/62494 , vital:28199
- Description: The NIST Cybersecurity Framework (CSF) is a specific risk and cybersecurity framework. It provides guidance on controls that can be implemented to help improve an organisation’s cybersecurity risk posture. The CSF Functions consist of Identify, Protect, Detect, Respond, and Recover. Like most Information Technology (IT) frameworks, there are elements of people, processes, and technology. The same elements are required to successfully implement the NIST CSF. This research specifically focuses on the technology element. While there are many commercial technologies available for a small to medium sized business, the costs can be prohibitively expensive. Therefore, this research investigates cost-effective technologies and assesses their alignment to the NIST CSF. The assessment was made against the NIST CSF subcategories. Each subcategory was analysed to identify where a technology would likely be required. The framework provides a list of Informative References. These Informative References were used to create high- level technology categories, as well as identify the technical controls against which the technologies were measured. The technologies tested were either open source or proprietary. All open source technologies tested were free to use, or have a free community edition. Proprietary technologies would be free to use, or considered generally available to most organisations, such as components contained within Microsoft platforms. The results from the experimentation demonstrated that there are multiple cost-effective technologies that can support the NIST CSF. Once all technologies were tested, the NIST CSF was extended. Two new columns were added, namely high-level technology category, and tested technology. The columns were populated with output from the research. This extended framework begins an initial collection of cost-effective technologies in support of the NIST CSF.
- Full Text:
- Date Issued: 2018
- Authors: Shackleton, Bruce Michael Stuart
- Date: 2018
- Subjects: National Institute of Standards and Technology (U.S.) , Computer security , Computer networks Security measures , Small business Information technology Cost effectiveness , Open source software
- Language: English
- Type: text , Thesis , Masters , MSc
- Identifier: http://hdl.handle.net/10962/62494 , vital:28199
- Description: The NIST Cybersecurity Framework (CSF) is a specific risk and cybersecurity framework. It provides guidance on controls that can be implemented to help improve an organisation’s cybersecurity risk posture. The CSF Functions consist of Identify, Protect, Detect, Respond, and Recover. Like most Information Technology (IT) frameworks, there are elements of people, processes, and technology. The same elements are required to successfully implement the NIST CSF. This research specifically focuses on the technology element. While there are many commercial technologies available for a small to medium sized business, the costs can be prohibitively expensive. Therefore, this research investigates cost-effective technologies and assesses their alignment to the NIST CSF. The assessment was made against the NIST CSF subcategories. Each subcategory was analysed to identify where a technology would likely be required. The framework provides a list of Informative References. These Informative References were used to create high- level technology categories, as well as identify the technical controls against which the technologies were measured. The technologies tested were either open source or proprietary. All open source technologies tested were free to use, or have a free community edition. Proprietary technologies would be free to use, or considered generally available to most organisations, such as components contained within Microsoft platforms. The results from the experimentation demonstrated that there are multiple cost-effective technologies that can support the NIST CSF. Once all technologies were tested, the NIST CSF was extended. Two new columns were added, namely high-level technology category, and tested technology. The columns were populated with output from the research. This extended framework begins an initial collection of cost-effective technologies in support of the NIST CSF.
- Full Text:
- Date Issued: 2018
Towards a framework for building security operation centers
- Authors: Jacobs, Pierre Conrad
- Date: 2015
- Subjects: Security systems industry , Systems engineering , Expert systems (Computer science) , COBIT (Information technology management standard) , Computer security
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4710 , http://hdl.handle.net/10962/d1017932
- Description: In this thesis a framework for Security Operation Centers (SOCs) is proposed. It was developed by utilising Systems Engineering best practices, combined with industry-accepted standards and frameworks, such as the TM Forum’s eTOM framework, CoBIT, ITIL, and ISO/IEC 27002:2005. This framework encompasses the design considerations, the operational considerations and the means to measure the effectiveness and efficiency of SOCs. The intent is to provide guidance to consumers on how to compare and measure the capabilities of SOCs provided by disparate service providers, and to provide service providers (internal and external) a framework to use when building and improving their offerings. The importance of providing a consistent, measureable and guaranteed service to customers is becoming more important, as there is an increased focus on holistic management of security. This has in turn resulted in an increased number of both internal and managed service provider solutions. While some frameworks exist for designing, building and operating specific security technologies used within SOCs, we did not find any comprehensive framework for designing, building and managing SOCs. Consequently, consumers of SOCs do not enjoy a constant experience from vendors, and may experience inconsistent services from geographically dispersed offerings provided by the same vendor.
- Full Text:
- Date Issued: 2015
- Authors: Jacobs, Pierre Conrad
- Date: 2015
- Subjects: Security systems industry , Systems engineering , Expert systems (Computer science) , COBIT (Information technology management standard) , Computer security
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4710 , http://hdl.handle.net/10962/d1017932
- Description: In this thesis a framework for Security Operation Centers (SOCs) is proposed. It was developed by utilising Systems Engineering best practices, combined with industry-accepted standards and frameworks, such as the TM Forum’s eTOM framework, CoBIT, ITIL, and ISO/IEC 27002:2005. This framework encompasses the design considerations, the operational considerations and the means to measure the effectiveness and efficiency of SOCs. The intent is to provide guidance to consumers on how to compare and measure the capabilities of SOCs provided by disparate service providers, and to provide service providers (internal and external) a framework to use when building and improving their offerings. The importance of providing a consistent, measureable and guaranteed service to customers is becoming more important, as there is an increased focus on holistic management of security. This has in turn resulted in an increased number of both internal and managed service provider solutions. While some frameworks exist for designing, building and operating specific security technologies used within SOCs, we did not find any comprehensive framework for designing, building and managing SOCs. Consequently, consumers of SOCs do not enjoy a constant experience from vendors, and may experience inconsistent services from geographically dispersed offerings provided by the same vendor.
- Full Text:
- Date Issued: 2015
Towards a portable occam
- Authors: Hill, David Timothy
- Date: 1988 , 2013-03-07
- Subjects: occam (Computer program language) , Transputers , Parallel programming (Computer science)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4587 , http://hdl.handle.net/10962/d1004749 , occam (Computer program language) , Transputers , Parallel programming (Computer science)
- Description: Occam is designed for concurrent programming on a network of transputers. AIlocation and partitioning of the program is specified within the source code, binding the program to a specific network. An altemative approach is proposed which completely separates the source code from hardware considerations. Static allocation is performed as a separate phase and should, ideally, be automatic but at present is manual. Complete hardware abstraction requires that non-local, shared communication be provided for, introducing an efficiency overhead which can be minimised by the allocation. The proposal was implemented on a network of IBM PCs, modelled on a transputer network, and implementation issues are discussed
- Full Text:
- Date Issued: 1988
- Authors: Hill, David Timothy
- Date: 1988 , 2013-03-07
- Subjects: occam (Computer program language) , Transputers , Parallel programming (Computer science)
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4587 , http://hdl.handle.net/10962/d1004749 , occam (Computer program language) , Transputers , Parallel programming (Computer science)
- Description: Occam is designed for concurrent programming on a network of transputers. AIlocation and partitioning of the program is specified within the source code, binding the program to a specific network. An altemative approach is proposed which completely separates the source code from hardware considerations. Static allocation is performed as a separate phase and should, ideally, be automatic but at present is manual. Complete hardware abstraction requires that non-local, shared communication be provided for, introducing an efficiency overhead which can be minimised by the allocation. The proposal was implemented on a network of IBM PCs, modelled on a transputer network, and implementation issues are discussed
- Full Text:
- Date Issued: 1988
Towards an evaluation and protection strategy for critical infrastructure
- Authors: Gottschalk, Jason Howard
- Date: 2015
- Subjects: Computer crimes -- Prevention , Computer networks -- Security measures , Computer crimes -- Law and legislation -- South Africa , Public works -- Security measures
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4721 , http://hdl.handle.net/10962/d1018793
- Description: Critical Infrastructure is often overlooked from an Information Security perspective as being of high importance to protect which may result in Critical Infrastructure being at risk to Cyber related attacks with potential dire consequences. Furthermore, what is considered Critical Infrastructure is often a complex discussion, with varying opinions across audiences. Traditional Critical Infrastructure included power stations, water, sewage pump stations, gas pipe lines, power grids and a new entrant, the “internet of things”. This list is not complete and a constant challenge exists in identifying Critical Infrastructure and its interdependencies. The purpose of this research is to highlight the importance of protecting Critical Infrastructure as well as proposing a high level framework aiding in the identification and securing of Critical Infrastructure. To achieve this, key case studies involving Cyber crime and Cyber warfare, as well as the identification of attack vectors and impact on against Critical Infrastructure (as applicable to Critical Infrastructure where possible), were identified and discussed. Furthermore industry related material was researched as to identify key controls that would aid in protecting Critical Infrastructure. The identification of initiatives that countries were pursuing, that would aid in the protection of Critical Infrastructure, were identified and discussed. Research was conducted into the various standards, frameworks and methodologies available to aid in the identification, remediation and ultimately the protection of Critical Infrastructure. A key output of the research was the development of a hybrid approach to identifying Critical Infrastructure, associated vulnerabilities and an approach for remediation with specific metrics (based on the research performed). The conclusion based on the research is that there is often a need and a requirement to identify and protect Critical Infrastructure however this is usually initiated or driven by non-owners of Critical Infrastructure (Governments, governing bodies, standards bodies and security consultants). Furthermore where there are active initiative by owners very often the suggested approaches are very high level in nature with little direct guidance available for very immature environments.
- Full Text:
- Date Issued: 2015
- Authors: Gottschalk, Jason Howard
- Date: 2015
- Subjects: Computer crimes -- Prevention , Computer networks -- Security measures , Computer crimes -- Law and legislation -- South Africa , Public works -- Security measures
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4721 , http://hdl.handle.net/10962/d1018793
- Description: Critical Infrastructure is often overlooked from an Information Security perspective as being of high importance to protect which may result in Critical Infrastructure being at risk to Cyber related attacks with potential dire consequences. Furthermore, what is considered Critical Infrastructure is often a complex discussion, with varying opinions across audiences. Traditional Critical Infrastructure included power stations, water, sewage pump stations, gas pipe lines, power grids and a new entrant, the “internet of things”. This list is not complete and a constant challenge exists in identifying Critical Infrastructure and its interdependencies. The purpose of this research is to highlight the importance of protecting Critical Infrastructure as well as proposing a high level framework aiding in the identification and securing of Critical Infrastructure. To achieve this, key case studies involving Cyber crime and Cyber warfare, as well as the identification of attack vectors and impact on against Critical Infrastructure (as applicable to Critical Infrastructure where possible), were identified and discussed. Furthermore industry related material was researched as to identify key controls that would aid in protecting Critical Infrastructure. The identification of initiatives that countries were pursuing, that would aid in the protection of Critical Infrastructure, were identified and discussed. Research was conducted into the various standards, frameworks and methodologies available to aid in the identification, remediation and ultimately the protection of Critical Infrastructure. A key output of the research was the development of a hybrid approach to identifying Critical Infrastructure, associated vulnerabilities and an approach for remediation with specific metrics (based on the research performed). The conclusion based on the research is that there is often a need and a requirement to identify and protect Critical Infrastructure however this is usually initiated or driven by non-owners of Critical Infrastructure (Governments, governing bodies, standards bodies and security consultants). Furthermore where there are active initiative by owners very often the suggested approaches are very high level in nature with little direct guidance available for very immature environments.
- Full Text:
- Date Issued: 2015
Towards large scale software based network routing simulation
- Authors: Herbert, Alan
- Date: 2015
- Subjects: Routers (Computer networks) , Computer software , Linux
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4709 , http://hdl.handle.net/10962/d1017931
- Description: Software based routing simulators suffer from large simulation host requirements and are prone to slow downs because of resource limitations, as well as context switching due to user space to kernel space requests. Furthermore, hardware based simulations do not scale with the passing of time as their available resources are set at the time of manufacture. This research aims to provide a software based, scalable solution to network simulation. It aims to achieve this by a Linux kernel-based solution, through insertion of a custom kernel module. This will reduce the number of context switches by eliminating the user space context requirement, and serve to be highly compatible with any host that can run the Linux kernel. Through careful consideration in data structure choice and software component design, this routing simulator achieved results of over 7 Gbps of throughput over multiple simulated node hops on consumer hardware. Alongside this throughput, this routing simulator also brings to light scalability and the ability to instantiate and simulate networks in excess of 1 million routing nodes within 1 GB of system memory
- Full Text:
- Date Issued: 2015
- Authors: Herbert, Alan
- Date: 2015
- Subjects: Routers (Computer networks) , Computer software , Linux
- Language: English
- Type: Thesis , Masters , MSc
- Identifier: vital:4709 , http://hdl.handle.net/10962/d1017931
- Description: Software based routing simulators suffer from large simulation host requirements and are prone to slow downs because of resource limitations, as well as context switching due to user space to kernel space requests. Furthermore, hardware based simulations do not scale with the passing of time as their available resources are set at the time of manufacture. This research aims to provide a software based, scalable solution to network simulation. It aims to achieve this by a Linux kernel-based solution, through insertion of a custom kernel module. This will reduce the number of context switches by eliminating the user space context requirement, and serve to be highly compatible with any host that can run the Linux kernel. Through careful consideration in data structure choice and software component design, this routing simulator achieved results of over 7 Gbps of throughput over multiple simulated node hops on consumer hardware. Alongside this throughput, this routing simulator also brings to light scalability and the ability to instantiate and simulate networks in excess of 1 million routing nodes within 1 GB of system memory
- Full Text:
- Date Issued: 2015