Skip to content

Commit

Permalink
removed the DOI form biblio
Browse files Browse the repository at this point in the history
  • Loading branch information
cb-unimi committed Jun 13, 2024
1 parent de00f9b commit da78f85
Show file tree
Hide file tree
Showing 2 changed files with 7 additions and 33 deletions.
27 changes: 1 addition & 26 deletions bib_on_BigDataAccessControl.bib
Original file line number Diff line number Diff line change
Expand Up @@ -122,7 +122,6 @@ @inproceedings{BigDataQaulitySurvey
volume = {},
number = {},
pages = {166--173},
doi = {10.1109/BigDataCongress.2018.00029}
}
@inproceedings{AConGraphDB:2021,
title = {{R}ole-{B}ased {A}ccess {C}ontrol on {G}raph {D}atabases},
Expand All @@ -142,7 +141,6 @@ @article{AConMongoDB:2022
volume = {},
number = {},
pages = {1--15},
doi = {10.1109/TETC.2022.3193577}
}
@inproceedings{HadoopStreaming:2020,
title = {{HBD}-{A}uthority: {S}treaming {A}ccess {C}ontrol {M}odel for {H}adoop},
Expand All @@ -161,7 +159,6 @@ @inproceedings{IoTSecurity
volume = {},
number = {},
pages = {1711--1715},
doi = {10.1109/TrustCom/BigDataSE.2018.00257}
}
@inproceedings{GuptaSandu:2017,
title = {{O}bject-{T}agged {RBAC} {M}odel for the {H}adoop {E}cosystem},
Expand All @@ -183,7 +180,6 @@ @inproceedings{MultipartyAC:2019
volume = {},
number = {},
pages = {29--38},
doi = {10.1109/EuroSPW.2019.00010}
}
@article{vassiliadis2009survey,
title = {A survey of extract--transform--load technology},
Expand Down Expand Up @@ -653,8 +649,6 @@ @inproceedings{Fuglede
author = {Fuglede, B. and Topsoe, F.},
booktitle = {International Symposium onInformation Theory, 2004. ISIT 2004. Proceedings.},
publisher = {IEEE},
doi = {10.1109/isit.2004.1365067},
url = {http://dx.doi.org/10.1109/ISIT.2004.1365067}
}
%%%%%%% FROM MEDES2 1 %%%%%%%
@article{castano2017exploratory,
Expand Down Expand Up @@ -692,9 +686,6 @@ @article{RAHMAN20102707
volume = 1,
number = 1,
pages = {2707--2716},
doi = {https://doi.org/10.1016/j.procs.2010.04.304},
issn = {1877-0509},
url = {https://www.sciencedirect.com/science/article/pii/S1877050910003054},
note = {ICCS 2010},
keywords = {Availability, Data mining, Forecasting, Enterprise grid, Jaccard Index},
abstract = {Enterprise Grid enables sharing and aggregation of a set of computing or storage resources connected by enterprise network, but the availability of the resources in this environment varies widely. Thus accurate prediction of the availability of these resources can significantly improve the performance of executing compute-intensive complex scientific and business applications in enterprise Grid environment by avoiding possible runtime failures. In this paper, we propose a Jaccard Index based prediction approach utilizing lazy learning algorithm that searches for a best match of a sequence pattern in the historical data in order to predict the availability of a particular machine in the system. We compare it against three other well known availability prediction techniques using simulation based study. The experimental results show that our Jaccard Index based prediction approach achieves better prediction accuracy with reduced computational complexity when compared to other similar techniques.}
Expand All @@ -709,7 +700,6 @@ @INPROCEEDINGS{7014544
number={},
pages={1-7},
keywords={Data structures;Boolean functions;Cryptography;Databases;Access Control;Authorization;Big Data;Distributed System},
doi={10.4108/icst.collaboratecom.2014.257649}
}

@InProceedings{dataProtection,
Expand Down Expand Up @@ -748,7 +738,7 @@ @ARTICLE{8863330
number={},
pages={147420-147452},
keywords={Cloud computing;Access control;Encryption;Privacy;Data privacy;Edge computing;Cloud computing;privacy;access control;attribute-based encryption;trust},
doi={10.1109/ACCESS.2019.2946185}}
}

@article{Majeed2021AnonymizationTF,
title={Anonymization Techniques for Privacy Preserving Data Publishing: A Comprehensive Survey},
Expand All @@ -769,7 +759,6 @@ @article{dataAccuracy
pages = {5--33},
year = {1996},
publisher = {Routledge},
doi = {10.1080/07421222.1996.11518099},
}

@article{dataQuality,
Expand All @@ -782,8 +771,6 @@ @article{dataQuality
volume = {41},
number = {2},
issn = {0001-0782},
url = {https://doi.org/10.1145/269012.269021},
doi = {10.1145/269012.269021},
journal = {Commun. ACM},
month = {feb},
pages = {54–57},
Expand All @@ -795,7 +782,6 @@ @Article{impetusPaper
title = "{Balancing Protection and Quality in Big Data Analytics Pipelines}",
journal = "Big Data",
year = "2024",
doi = {10.1089/big.2023.0065},
}

@book{bookMetrics,
Expand Down Expand Up @@ -851,9 +837,6 @@ @article{VANDENBROEK2018330
volume = {129},
pages = {330-338},
year = {2018},
issn = {0040-1625},
doi = {https://doi.org/10.1016/j.techfore.2017.09.040},
url = {https://www.sciencedirect.com/science/article/pii/S0040162517314695},
author = {Tijs {van den Broek} and Anne Fleur {van Veenstra}},
keywords = {Disruptive innovation, Data protection regulation, Big data, Governance, Inter-organizational collaboration},
abstract = {Big data is an important driver of disruptive innovation that may increase organizations' competitive advantage. To create innovative data combinations and decrease investments, big data is often shared among organizations, crossing organizational boundaries. However, these big data collaborations need to balance disruptive innovation and compliance to a strict data protection regime in the EU. This paper investigates how inter-organizational big data collaborations arrange and govern their activities in the context of this dilemma. We conceptualize big data as inter-organizational systems and build on IS and Organization Theory literature to develop four archetypical governance arrangements: Market, Hierarchy, Bazaar and Network. Subsequently, these arrangements are investigated in four big data collaboration use cases. The contributions of this study to literature are threefold. First, we conceptualize the organization behind big data collaborations as IOS governance. Second, we show that the choice for an inter-organizational governance arrangement highly depends on the institutional pressure from regulation and the type of data that is shared. In this way, we contribute to the limited body of research on the antecedents of IOS governance. Last, we highlight with four use cases how the principles of big data, specifically data maximization, clash with the principles of EU data protection regulation. Practically, our study provides guidelines for IT and innovation managers how to arrange and govern the sharing of data among multiple organizations.}
Expand All @@ -867,7 +850,6 @@ @article{needtobalance
title = {Balancing data privacy and usability in the federal statistical system},
volume = {119},
journal = {Proceedings of the National Academy of Sciences},
doi = {10.1073/pnas.2104906119}
}

@INPROCEEDINGS{secureWScomposition,
Expand All @@ -879,7 +861,6 @@ @INPROCEEDINGS{secureWScomposition
number={},
pages={489-496},
keywords={Security;Web services;XML;Simple object access protocol;Service oriented architecture;Privacy;Software systems;Software design;Web and internet services;Computer architecture},
doi={10.1109/ICWS.2006.115}
}

@article{SELLAMI2020102732,
Expand All @@ -889,8 +870,6 @@ @article{SELLAMI2020102732
pages = {102732},
year = {2020},
issn = {1084-8045},
doi = {https://doi.org/10.1016/j.jnca.2020.102732},
url = {https://www.sciencedirect.com/science/article/pii/S108480452030206X},
author = {Mokhtar Sellami and Haithem Mezni and Mohand Said Hacid},
keywords = {Big data, Big service, Big service composition, Quality of big services, Fuzzy RCA, Spark},
abstract = {Over the last years, big data has emerged as a new paradigm for the processing and analysis of massive volumes of data. Big data processing has been combined with service and cloud computing, leading to a new class of services called “Big Services”. In this new model, services can be seen as an abstract layer that hides the complexity of the processed big data. To meet users' complex and heterogeneous needs in the era of big data, service reuse is a natural and efficient means that helps orchestrating available services' operations, to provide customer on-demand big services. However different from traditional Web service composition, composing big services refers to the reuse of, not only existing high-quality services, but also high-quality data sources, while taking into account their security constraints (e.g., data provenance, threat level and data leakage). Moreover, composing heterogeneous and large-scale data-centric services faces several challenges, apart from security risks, such as the big services' high execution time and the incompatibility between providers' policies across multiple domains and clouds. Aiming to solve the above issues, we propose a scalable approach for big service composition, which considers not only the quality of reused services (QoS), but also the quality of their consumed data sources (QoD). Since the correct representation of big services requirements is the first step towards an effective composition, we first propose a quality model for big services and we quantify the data breaches using L-Severity metrics. Then to facilitate processing and mining big services' related information during composition, we exploit the strong mathematical foundation of fuzzy Relational Concept Analysis (fuzzy RCA) to build the big services' repository as a lattice family. We also used fuzzy RCA to cluster services and data sources based on various criteria, including their quality levels, their domains, and the relationships between them. Finally, we define algorithms that parse the lattice family to select and compose high-quality and secure big services in a parallel fashion. The proposed method, which is implemented on top of Spark big data framework, is compared with two existing approaches, and experimental studies proved the effectiveness of our big service composition approach in terms of QoD-aware composition, scalability, and security breaches.}
Expand All @@ -905,7 +884,6 @@ @ARTICLE{9844845
number={3},
pages={1999-2012},
keywords={Software;Behavioral sciences;Microservice architectures;Art;Monitoring;Focusing;Codes;Assurance;certification;security;service selection},
doi={10.1109/TSC.2022.3195071}
}

@InProceedings{Lseverity,
Expand Down Expand Up @@ -957,9 +935,6 @@ @Inbook{Kellerer2004
address="Berlin, Heidelberg",
pages="235--283",
abstract="In this first chapter of extensions and generalizations of the basic knapsack problem (KP) we will add additional constraints to the single weight constraint (1.2) thus attaining the multidimensional knapsack problem. After the introduction we will deal extensively with relaxations and reductions in Section 9.2. Exact algorithms to compute optimal solutions will be covered in Section 9.3 followed by results on approximation in Section 9.4. A detailed treatment of heuristic methods will be given in Section 9.5. Separate sections are devoted to two special cases, namely the two-dimensional knapsack problem (Section 9.6) and the cardinality constrained knapsack problem (Section 9.7). Finally, we will consider the combination of multiple constraints and multiple-choice selection of items from classes (see Chapter 11 for the one-dimensional case) in Section 9.8.",
isbn="978-3-540-24777-7",
doi="10.1007/978-3-540-24777-7_9",
url="https://doi.org/10.1007/978-3-540-24777-7_9"
}


13 changes: 6 additions & 7 deletions metrics.tex
Original file line number Diff line number Diff line change
Expand Up @@ -62,13 +62,12 @@ \subsubsection{Pipeline Quality}
%We also note that information loss \textit{dloss} is used to generate the Max-Quality pipeline instance in the remaining of this section.

\subsection{NP-Hardness of the Max-Quality Pipeline Instantiation Problem}\label{sec:nphard}
%\hl{se lo definiamo in maniera formale come il problema di trovare un'istanza valida in accordo alla definizione di istanza tale che non ne esiste una con un loss piu' piccolo?}
The problem of computing a pipeline instance (\cref{def:instance}) with maximum quality \q\ can be formally defined as follows.

%The problem of computing a pipeline instance (\cref{def:instance}) with maximum quality \q\ can be formally defined as follows.
The process of computing a pipeline instance (\cref{def:instance}) with maximum quality \q\ can be formally defined as follows.
\vspace{0.5em}

\begin{definition}[Max-Quality Problem]\label{def:MaXQualityInstance}
Given a pipeline template $G^{\myLambda,\myGamma}$ and a set $S^c$ of candidate services, find a max-quality pipeline instance $G'$ such that:
Given a pipeline template $G^{\myLambda,\myGamma}$ and a set $S^c$ of candidate services, find a max-quality pipeline instance $G'$ such that:
\begin{itemize}
\item $G'$ satisfies conditions in \cref{def:instance},
\item $\nexists$ a pipeline instance $G''$ that satisfies conditions in \cref{def:instance} and such that quality \q($G''$)$>$\q($G'$), where \q($\cdot$) is the pipeline quality in Definition~\ref{def:quality}.
Expand All @@ -78,19 +77,19 @@ \subsection{NP-Hardness of the Max-Quality Pipeline Instantiation Problem}\label

\vspace{0.5em}

The Max-Quality \problem is a combinatorial selection problem and is NP-hard, as stated by \cref{theorem:NP}. However, while the overall problem is NP-hard, the filtering step of the process, is solvable in polynomial time.
The Max-Quality problem is a combinatorial selection problem and is NP-hard, as stated by \cref{theorem:NP}. However, while the overall problem is NP-hard, the filtering step of the process, is solvable in polynomial time.
%However, while the overall problem is NP-hard, there is a component of the problem, i.e., matching the profile of each service with the corresponding vertex policy, that is solvable in polynomial time.
It can be done by iterating over each vertex and each service, checking if the service matches the vertex policy. This process takes polynomial time complexity $O(|N|*|S|)$.

\vspace{0.5em}

\begin{theorem}\label{theorem:NP}
The Max-Quality \problem is NP-Hard.
The Max-Quality problem is NP-Hard.
\end{theorem}
\emph{Proof: }
The proof is a reduction from the multiple-choice knapsack problem (MCKP), a classified NP-hard combinatorial optimization problem, which is a generalization of the simple knapsack problem (KP) \cite{Kellerer2004}. In the MCKP problem, there are $t$ mutually disjoint classes $N_1,N_2,\ldots,N_t$ of items to pack in some knapsack of capacity $C$, class $N_i$ having size $n_i$. Each item $j$$\in$$N_i$ has a profit $p_{ij}$ and a weight $w_{ij}$; the problem is to choose one item from each class such that the profit sum is maximized without having the weight sum to exceed C.

The MCKP can be reduced to the Max quality \problem in polynomial time, with $N_1,N_2,\ldots,N_t$ corresponding to the sets of compatible services $S^c_{1}, S^c_{2}, \ldots, S^c_{u}$, with $t$$=$$u$ and $n_i$ also the size of each set $S^c_{i}$. The profit $p_{ij}$ of item $j$$\in$$N_i$ corresponds to quality \textit{\q}$_{ij}$ computed for each candidate service $s_j$$\in$$S^c_{i}$, while $w_{ij}$ is uniformly 1 (thus, C is always equal to the cardinality of $V_C$). It is evident that the solution to one problem is also the solution to the other (and vice versa). Since the reduction can be done in polynomial time, the Max-Quality \problem is also NP-hard.
The MCKP can be reduced to the Max quality \problem in polynomial time, with $N_1,N_2,\ldots,N_t$ corresponding to the sets of compatible services $S^c_{1}, S^c_{2}, \ldots, S^c_{u}$, with $t$$=$$u$ and $n_i$ also the size of each set $S^c_{i}$. The profit $p_{ij}$ of item $j$$\in$$N_i$ corresponds to quality \textit{\q}$_{ij}$ computed for each candidate service $s_j$$\in$$S^c_{i}$, while $w_{ij}$ is uniformly 1 (thus, C is always equal to the cardinality of $V_C$). It is evident that the solution to one problem is also the solution to the other (and vice versa). Since the reduction can be done in polynomial time, the Max-Quality problem is also NP-hard.

\vspace{0.5em}

Expand Down

0 comments on commit da78f85

Please sign in to comment.