Masterarbeit, 2017
40 Seiten, Note: 10
1. INTRODUCTION
1.1 IDEA AND MOTIVATION
1.2 LITERATURE SURVEY
2. PROBLEM DEFINITION AND SCOPE
2.1 SCOPE
2.2 SOFTWARE CONTEXT
2.3 SOFTWARE CONSTRAINTS
2.4 OUTCOMES
2.5 HARDWARE SPECIFICATION
2.6 S/W SPECIFICATION
2.7 AREA OF DISSERTATION
3. DISSERTATION PLAN
3.1 PROJECT PLAN
3.2 TIMELINE OF PROJECT
3.3 FEASIBILITY STUDY
3.3.1 Economical Feasibility
3.3.2 Technical Feasibility
3.3.3 Operational feasibility
3.3.4 Time Feasibility
3.4 RISK MANAGEMENT
3.4.1 Project Risk:
3.4.2 Risk Assessment
3.5 EFFORT AND COST ESTIMATION
3.5.1 Lines of code (LOC)
3.5.2 Effort
3.5.3 Development Time
3.5.4 Number of People
4. SOFTWARE REQUIREMENT SPECIFICATION
4.1 INTRODUCTION
4.1.1 Purpose
4.1.2 Scope of Document
4.1.3 Overview of responsibilities of developer
4.2 PRODUCT OVERVIEW
4.2.1 Block diagram
4.3 FUNCTINAL MODEL
4.3.1 Flow diagram
4.3.2 Data Flow Diagram
4.3.3 UML Diagrams
4.3.3.1 Sequence diagram
4.3.3.2 Class diagram
4.3.4 Non-Functional Requirements
4.4 BEHAVIORAL MODEL AND DESCRIPTION
4.4.1 Description of software behavior
4.4.2 Use case diagram
5. DETAILED DESIGN
5.1 ARCHITECTURE DESIGN
5.1.1 Algorithms:
5.2 INTERFACES
5.2.1 Human Interface
5.2.2. Database interface
6. TESTING
6.1 INTRODUCTION
6.1.1 Goals and Objective
6.2 TESTING STRATEGY
6.2.1 White Box Testing
6.2.2 Black Box Testing
6.2.3 System testing
6.2.4 Performance testing
7. DATA TABLE AND DISCUSSION
7.1 INPUT TO THE SYSTEM
7.2 OUTPUT:
7.3 PERFORMANCE OF PROPOSED SYSTEM
7.3.1 Performance of proposed system with respect to baseline algorithm:
7.3.2 Performance of proposed system with respect to blowfish encryption algorithm:
7.4 RESULT
7.4.1 Difference between proposed algorithm and base algorithm i.e provider aware algorithm:
8. SUMMARY AND CONCLUSION
8.1 FUTURE ENHANCEMENT
The primary objective of this work is to develop an effective data mining and anonymization method to protect sensitive information during collaborative data publishing in distributed network environments, aiming to achieve higher performance compared to existing encryption-based approaches.
5.1.1 Algorithms:
Slicing is basically depends on Attribute and tuple partitioning. In Attribute partitioning (vertical partition) we partitioned data as {name},{age-zip} and {Disease} and tuple partitioning (horizontal partition) as {t1,t2,t3,t4,t5,t6}. In attribute partitioning age and zip are partitioned together because they both are highly correlated because they are quasi identifiers (QI). These QI can be known to attacker. While tuple partitioning system should check L diversity for the sensitive attribute (SA) column. Algorithm runs are as follows.
1. Initialize bucket k=n, int i= rowcount, column count=C, Q={D}, // D= data into database, Arraylist= a[i];
2. While Q is not empty
If i<=n
Check L diversity;
Else
i++;
Return D*;
3. Q=Q-{D*+a[i]};
4. Repeat step 2 and 3 with next tuple in Q
5. D*=D*U A[D] // next anonymized view of data D
1. INTRODUCTION: Discusses the necessity of privacy-preserving data sharing in distributed environments and introduces the motivation behind using slicing techniques over traditional encryption models.
2. PROBLEM DEFINITION AND SCOPE: Defines the goal of creating an anonymized, attack-immune data view and outlines the software context, constraints, and hardware/software environment.
3. DISSERTATION PLAN: Provides a comprehensive project timeline, feasibility study (economical, technical, operational, and time), and detailed risk assessment and cost estimation metrics.
4. SOFTWARE REQUIREMENT SPECIFICATION: Details the functional and behavioral models, including block diagrams, data flow diagrams, and UML diagrams for the proposed system.
5. DETAILED DESIGN: Explains the architectural design and the specific slicing and L-diversity algorithms implemented to maintain data privacy.
6. TESTING: Describes the testing strategies employed, including white box, black box, and performance testing, to validate system reliability and computation efficiency.
7. DATA TABLE AND DISCUSSION: Evaluates the system results by comparing the computation time and performance complexity of the proposed slicing algorithm versus baseline and blowfish encryption models.
8. SUMMARY AND CONCLUSION: Summarizes the research findings regarding privacy protection in collaborative data publishing and suggests potential future improvements, such as ad hoc grid computing implementation.
Data Mining, Slicing, Anonymization, L-diversity, Distributed Database, Privacy Preserving, Blowfish Encryption, Collaborative Data Publishing, Insider Attack, Quasi Identifiers, Sensitive Attribute, Security, Software Requirement Specification, Data Utility, Performance Metrics
The research focuses on enhancing data privacy during collaborative data publishing across distributed databases by utilizing slicing techniques as an alternative to computationally expensive encryption methods.
Key areas include data anonymization, attribute and tuple partitioning, performance measurement of algorithms, the L-diversity model, and security assurance in distributed networks.
The aim is to provide a privacy-preserved, anonymized view of integrated data from different providers that is immune to attacks while maintaining optimal computation time.
The work employs a comparative methodology, using algorithmic analysis and performance evaluation (computation time and complexity) to contrast the proposed Slicing method with existing encryption and provider-aware algorithms.
The main body focuses on the design of the system, the development of the Slicing and L-diversity algorithms, and the rigorous testing (black/white box) of these implementations against various threat scenarios.
The work is best characterized by terms such as Data Mining, Slicing, L-diversity, Distributed Database, Anonymization, and Privacy Preserving.
Yes, the system is designed to detect and protect against "insider attacks" where entities may attempt to breach privacy using background knowledge.
The proposed method offers a significant reduction in computation time, making it more efficient for systems like hospital patient data management or banking where rapid access is required alongside privacy.
Validation is performed through a multi-stage testing process including white-box and black-box testing, alongside graphical performance comparisons illustrating CPU usage and execution time.
Der GRIN Verlag hat sich seit 1998 auf die Veröffentlichung akademischer eBooks und Bücher spezialisiert. Der GRIN Verlag steht damit als erstes Unternehmen für User Generated Quality Content. Die Verlagsseiten GRIN.com, Hausarbeiten.de und Diplomarbeiten24 bieten für Hochschullehrer, Absolventen und Studenten die ideale Plattform, wissenschaftliche Texte wie Hausarbeiten, Referate, Bachelorarbeiten, Masterarbeiten, Diplomarbeiten, Dissertationen und wissenschaftliche Aufsätze einem breiten Publikum zu präsentieren.
Kostenfreie Veröffentlichung: Hausarbeit, Bachelorarbeit, Diplomarbeit, Dissertation, Masterarbeit, Interpretation oder Referat jetzt veröffentlichen!

