In the most elementary sense, the research design is the logical sequence that connects the empirical data to a study’s initial research questions and, ultimately, to its conclusions (Yin, 2003). It can also be seen as a blueprint, chain of evidence, or logical model of proof. It needs to maximize construct validity, internal validity, external validity and reliability.
In this research design, I present the methodological issues of the thesis. It presents the unit of analysis, it covers the reasons for selecting organizations, it describes the data sources that where used, how the data was collected and how the data was analyzed. It is presented in such a way that other researchers can replicate this research (Yin, 2003).
Unit of analysis
For a case study it is important to define the case, in terms of what the case is, and where the case leaves off (Miles and Huberman, 1994; Yin, 2003). This is a problem for many researchers in case studies. The more a study contains specific propositions, the more it will stay within feasible limits, and also the context has to be clear (Yin, 2003).
This case study is multiple and holistic (Yin, 2003). I have conducted interviews within two different organisations. Case studies can be single or multiple-case designs, where a multiple design must follow a replication rather than sampling logic. When no other cases are available for replication, the researcher is limited to single-case designs. One of the rationales to justify a single case study is if theory has specified a clear set of testable propositions (Yin, 2003), which is the case in this research. To test the propositions I made use of the customization characteristics identified in the literature study. Also theory has to specify the circumstances within which the propositions are believed to be true (Yin, 2003). This is also the case in this research, because it focuses on online delivered content as the unit of analysis. During this research, I had the opportunity to conduct multiple case studies. Multiple case studies are preferred, because they can be more robust than a single case study and, depending on the results, can strengthen the external validity (Yin, 2003).
Single case- and multiple case studies can further be classified as holistic or embedded. In an embedded case study, the case is split in multiple units of analysis, while a holistic case study has one unit of analysis for each case. The unit of analysis of each case is that part of the company that is relevant to answer the main research question, also called logical subunits (Yin, 2003). When no logical subunits can be identified, the holistic design is advantageous. When this unit of analysis changes during the study, the researcher can be forced to start over. When conducting an embedded case study, the researcher has the pitfall to focus too much on a single unit, and fails to return to the larger unit of analysis.
Site selection criteria
During this thesis it was not possible to collect data from many sources. The sites to study should be appropriate and accessible. There are simply not many sources available where digital products are being customized over the Internet, which is a requisite to answer the main research question and the hypotheses that where identified during the literature review. The site to study should also be a location where entry or access to the sources should be available, and the appropriate people should likely be available (Berg, 2004). The logic of using samples is to make inferences about some larger population from a smaller one, which is the sample (Berg, 2004).
There are various sampling principles to select the sites to study, such as maximum variation sampling, critical case sampling, snowball sampling, purposive sampling or convenience sampling (Miles and Huberman, 1994; Berg, 2004). Sampling involves decisions about which people to observe or interview. Maximum variation sampling, for example, involves looking for outliers to see whether main patterns still hold. Qualitative samples tend to be purposive rather than random, which is very important with small numbers of cases (Miles and Huberman, 1994). Two actions are involved when sampling in qualitative research. First, the boundaries have to be set to define aspects of the cases that can be studied within the limits of the available time and means, that connects directly to the research questions and that probably will include samples of what needs to be studied. Second, a sampling frame needs to be created to help to uncover, confirm or qualify the basic processes or constructs of the study (Miles and Huberman, 1994).
During this research I made use of the principle of multiple-case sampling. Multiple case sampling adds confidence to findings. This approach connects directly to the overall research question, adding confidence on the ‘how’ research question. The overall research question:
How to support customization and personalization for pure digital products in the Internet economy to dramatically decrease complexity and search costs for consumers, so variety can be maximized?
When using this kind of sampling, an explicit sampling frame is needed (Miles and Huberman, 1994). The sampling frame was created by identifying suppliers that offer online digital products on the Internet, that can be customized, preferably have a large variety, and it should be easy for consumers to find the digital products they would like. Organizations that offer digital products where variety is high, and offer these products online are scarce; in particular companies that offer possibilities to customize their products. Nowadays, digital products in the form of music seem to qualify for these criteria. The two most popular organizations that offer customized music are Last.fm and Pandora Media. Both organizations agreed to be part of this research. These two sites offer digital products in the form of streaming music on the Internet. Another company who offers digital products in the form of music is Mercora, but they did not agree to participate.
Yin (2003) identifies six sources of evidence that can be collected during case studies, each having their strengths and weaknesses. The first is documentation, which is stable because it can be reviewed repeatedly, it is unobtrusive, it is exact and it has a broad coverage. However, it can also be difficult to retrieve, the selection and reporting can be biased, and the access can deliberately be blocked. The second is archival records, which is the same as documentation, but in addition it also has the advantage of being precise and quantitative, and the disadvantage of being difficult to access due to privacy reasons. The third is interviews. Interviews are targeted and insightful, but also have disadvantages because they can be biased due to poorly constructed questions or poor responses, and they can be inaccurate due to poor recall. The fourth are direct observations which have the advantage of being real-time and contextual, but they can be time consuming, selective, the observed event may react different due to the observation, and is time consuming. The fifth is participant-observation which has the same characteristics as direct observations. They have an extra advantage as being insightful into interpersonal behaviour and an extra disadvantage of the possibility being biased due to manipulation. The sixth and last source of evidence is physical artefacts. They are insightful into cultural features and technical operations; however, selectivity and availability are disadvantages.
When conducting a case study, three principles of data collection can maximize the benefits of the above six sources of evidence (Yin, 2003). The first is to use multiple sources of evidence, which, if done properly, enables data triangulation. It also helps to avoid tunnel vision (Verschuren, 2003). The second principle is to create a case study database. Yin (2003) recommends keeping the data or evidence and reports separated. The last principle is to maintain a chain of evidence, which increases the reliability of the information.
There are many possible sources of evidence to identify, for example documentation and archival records in the form of existing reports. Another source of evidence that was used was by means of interviews. The semi structured interview was used, because the topics are clear and some questions can be predetermined, but it leaves space for probing beyond given answers (Berg, 2004). The interviews were recorded, transcribed and reviewed by the interviewees. The remaining sources of evidence, which are direct observations, participant-observation, and physical artefacts, were not used.
Data analysis in qualitative research can be defined as consisting of three concurrent flows of action: data reduction, data display, and conclusions and verification. These flows are present in parallel during and after the collection of data (Miles and Huberman, 1994). Data reduction refers to the process of selecting, focusing, simplifying, abstracting and transforming the collected data. It needs to be reduced in order to make the data more readily accessible and understandable (Berg, 2004; Kvale, 1996). Data display is intended to organize the collected data in such a way that it permits conclusion drawing (Miles and Huberman, 1994; Berg, 2004). The third component of the data analysis process is conclusion drawing and verification. During the collection of data, there should not be made any definitive conclusions, and these preliminary conclusions should be verified during the process (Miles and Huberman, 1994).
Linking data to propositions can be done in a number of ways, for example the technique of pattern matching, whereby several pieces of the same case may be related to some theoretical proposition (Yin, 2003). Other strategies are explanation building, time-series analysis, logic models, and cross-case synthesis. Every case study should strive to have a general analytic strategy, defining priorities for what to analyze and why. Examples of such general strategies are relying on theoretical propositions, thinking about rival explanations, and developing a case description (Yin, 2003).
The general analytic strategy that I used for data analysis is the technique of relying on theoretical propositions. The four hypotheses that followed from reviewed literature and the overall research question, led to this case study. To develop internal validity and external validity, I followed the specific analytical technique of pattern matching (Yin, 2003). When all collected data is available in textual format, data can be methodologically analyzed (Miles and Huberman, 1994). In pattern matching, or pattern coding, an empirically based pattern is compared with a predicted or proposed one. Pattern coding has four important functions (Miles and Huberman, 1994). First, it reduces large amounts of data into a smaller number of analytical units. Second, it gets the researcher into analysis during data collection, so that later fieldwork can be more focused. Third, it helps the researcher elaborate a cognitive map for understanding interactions. Fourth, it lays the groundwork for cross-case analysis by surfacing common themes.
The above strategy was used as a strategy during the case study. One of the data sources, and also the most dominant, existed of transcribed interviews, as I conducted three interviews within the two companies during this research. Kvale (1996) differentiates between five main approaches to analyze interviews. These are meaning condensation, meaning categorization, meaning structuring through narratives, meaning interpretation, and ad hoc meaning generation. During this research I used a combination of the meaning condensation approach and the meaning categorization approach. Meaning condensation entails an abridgement of the meaning expressed by the interviewees into shorter formulations. Long statements are compressed into briefer statements in which the main sense of what is being said is rephrased in a few words. Meaning condensation thus involves a reduction of large interview texts into briefer more succinct formulations (Kvale, 1996). Meaning categorization implies that the interview is coded into categories. Long statements are reduced to predefined categories, and can reduce and structure a large text into a few tables and figures (Kvale, 1996).
This paragraph summarizes the instrumentation which is used for this research. All measures are based on the review of the literature in Chapter 2.
Mass customization classification
To classify the mass customizer in terms of consumer involvement in the design process and product modularity, I follow Duray et al. (2000). They developed an instrument to classify mass customizers, with established scales to enhance validity, reliability and generalizability of measures.
According to Duray et al. (2000), consumer involvement can be scaled into two factors. The first factor is consumer involvement in the design and fabrication stages, and is considered as a high degree of customization. The second factor is consumer involvement in the assembly and use stages, and is considered as a low degree of customization. To measure the type of modularity employed, Duray et al. (2000) also identified two factors. The first factor is modularity through fabrication, and can be considered a measure of modularity in the design or fabrication of a product. The second factor is modularity through standardization. It contains items that address modularity in the form of options to standard products or interchangeability of components.
To operationalize the concept of point of customer involvement, the earliest point of involvement classifies the company. Once a customer is involved in the process, involvement carries throughout the whole production cycle. If a customer’s initial point of involvement is in the design stage of the production cycle, the customer’s preferences would be incorporated throughout the remaining stages of fabrication, assembly and use (Duray et al., 2000). The same is the case for the type of modularity employed. Once each company has been assigned one value for each of the variables, customer involvement and modularity, the classification process is simplified. Table 3.1 shows the identification of mass customizers.
Design / fabrication
Assembly / use
Design / fabrication
Assembly / use
Table 3.1: Classification of the mass customization configuration.
To measure variety, I follow Blecker et al. (2006) who propose a key metrics based approach to control variety induced complexity in mass customization. Blecker et al. (2006) revealed that multiple use, interface complexity and platform efficiency are key metrics that directly influence the extent of product variations that can be offered by the mass customizer. The multiple use metric provides a measurement of the number of product variant required by consumers as compared to the total number of modules (Ericsson and Erixon, 1999 in Blecker et al., 2006).
Blecker et al., 2006
Blecker et al., 2006
Blecker et al., 2006
Table 3.2: Possible variety metrics.
Complexity and search costs
The literature review revealed that measuring complexity is not easy, because it is very subjective, and depends on how the consumption experience is experienced (Desmeules, 2002). To evaluate the extent of perceived complexity, I follow Blecker et al. (2006), who proposes two metrics. The first metric measures the average interaction length of time, in other words, how much time consumers need to on average to completely configure a product variant. The second key metric refers to the abortion rate. If consumers are uncertain about their choices or overwhelmed by the interaction process, it is more likely that they give up configuration and leave the website of the mass customizer.
Average interaction length of time
Blecker et al., 2006;
Blecker et al., 2006
Table 3.3: Perceived complexity metrics.
To answer the main research question, I will have to be able to identify if search costs are dramatically reduced for consumers. These identified measures are primarily taken from Kurniawan et al. (2006), who did a study about decision quality with product selection.
Number of alternatives searched
Kurniawan et al., 2006
Kurniawan et al., 2006
Helander and Khalid, 2000
Table 3.4: Search costs metrics.
To identify methods or strategies that the mass customizer uses to reduce the perceived complexity and search costs, I will use the following metrics:
Reduce perceived complexity (customization)
Attribute vs. alternative
Huffman and Kahn, 1998
Dellaert and Stremersch, 2004
Dellaert and Stremersch, 2004; Piller et al., 2005;
Huffman and Kahn, 1998
Stegmann et al., 2006
Stegmann et al., 2006
Piller et al., 2005
Piller et al., 2005
Table 3.5: Reduce perceived complexity, or customization metrics.
Case study protocol and case study database
The case studies are conducted following a protocol containing the following.
- Procedures to introduce the case to the interviewees;
- Procedures to start and finish a case study;
- Procedures for conducting interviews including initial questions;
- Procedures for data recording in the case study database.
For every case, a case study database is constructed with the following structure.
- Introducing e-mails to the interviewees;
- Recorded interviews;
- Literal transcription of the interviews;
- Downloaded documents from the companies’ websites;
- Earlier interviews found on the Internet.
An embedded case study is a case study containing more than one sub-unit of analysis (Yin, 2003). Similar to a case study, an embedded case study methodology provides a means of integrating quantitative and qualitative methods into a single research study (Scholz & Tietje, 2002; Yin 2003). However, the identification of sub-units allows for a more detailed level of inquiry. The embedded case study design is an empirical form of inquiry appropriate for descriptive studies, where the goal is to describe the features, context, and process of a phenomenon. Roland W. Scholz suggests that “case is faceted or embedded in a conceptual grid” which allows to identify key components of human and environmental systems (Scholz 2011, p. 25).
A case study research methodology relies on multiple sources of evidence to add breadth and depth to data collection, to assist in bringing a richness of data together in an apex of understanding through triangulation, and to contribute to the validity of the research (Yin, 2003). The unique strength of this approach is this ability to combine a variety of information sources including documentation, interviews, and artifacts (e.g., technology or tools).
"The case study is preferred in examining contemporary events, when the relevant behaviours cannot be manipulated" (Yin, 2003, p. 7). The embedded case study approach is particularly relevant to examination of an environment where the boundaries between the phenomenon of interest and context are not clearly evident.
- Scholz, R. W. & Tietje, O. (2002). Embedded Case Study Methods: Integrating Quantitative and Qualitative Knowledge. London: Sage Publications Inc. ISBN 0-7619-1946-5
- Yin, R. K. (2003). Case study research, design and methods, 3rd ed. Newbury Park: Sage Publications. ISBN 0-7619-2553-8