Paper Menu >>
Journal Menu >>
			![]() J. Serv. Sci. & Management. 2008, 1: 91-100  Published Online June 2008 in SciRes (www.SRPublishing.org/journal/jssm)  Copyright © 2008 SciRes                                                                                 JSSM  Website Design Quality and Form Input Validation: An  Empirical Study on Irish Corporate Websites  Mary Levis1, Markus Helfert1 and Malcolm Brady1  1 Dublin City University  ABSTRACT  The information ma intained ab out products, services  and cu stomers is a m ost valuable o rganisational  asset. Th erefore,  it is important for successful electronic business to have high quality websites. A website must however, do more than  just look attractive it must be usable and presen t useful, usable information. Usability essen tia lly means tha t the website   is intuitive and allows visitors to find what they are looking for quickly and without effort. This means careful consid- eration of the structure of information and navigational design. According to the Open Web Applications Security Pro- ject, invalidated input is one of the top ten critical web-application security vulnerabilities. We empirically tested 21  Irish corporate websites. The findings suggested that one of the biggest problems is that many failed to use mechanisms  to validate even the basic user data input at the source of collection which could potentially result in a database full of  useless information.   Keywords: Website Design Quality, Form Input Validation, Information Quality, Data Quality  1. Introduction  The World Wide Web (WWW) is the largest available  distributed dynamic repository of information, and has  undergone massive and rapid growth since its inception.  There are over 2,060,000 users in Ireland alone. Over the  last seven years (2000 - 2007), Internet usage in Ireland  has grown by 162.8%; in United Kingdom by 144.2%; in  Europe by 221. 5% an d Wo rl d wi de by  24 4.7% [18].  Based on these facts, the Internet has assumed a central  role in many aspects of our lives and therefore creates a  greater need for businesses to design better websites in  order to stay competitive and increase revenue. Interac- tivity is essential to engage visitors and lead them to the  desired action and customers are more likely to return to a  website that has useful interactivity.   The website's homepage should be a marketing tool  designed as a  'bil lboard'  for t he or ganizat ion. The  desi gn is  critical in capturing the viewer's attention and interest [25]  and should represent the company in a meaningful and  positive light.  Therefore, there are many web design  concerns for commercial organizations when designing  their website.   The most basic are as follows: content that should be  included, selecting relevant and essential information,  designing a secure, usable, user friendly web interface that  is relatively easy to nav igate, and ensuring the site is easy  to find us ing any of the ma jor se arch eng ines. In  the driv e  to make the website look appealing from a visual per- spective other factors are often ignored, such as validation  and security, which leads to poor user experience and data  quality problems.   Data in the real world is constantly changing therefore  feedback is necessary in order to ensure that quality is  maintained. Data is dee med of high quality if it ‘correctly  represents the real-world construct to which it refers so  that products or decisions can be made’ [30]. One can  probably find as many definitions for quality on the web as  there are papers on quality. There are however, a number  of theoretical frameworks for understanding data quality.   Redman [33] and Orr [27] have presented cybernetic  models of information quality. The cybernetic view con- siders organizations as made up of closely interacting  feedback systems that link quality of information to how it  is used, in a feedback cycle where the actions of each  system is continuously modified by the actions, changes  and outputs of the others [2,29,36]. Figure 1 shows an  information system in the real world context.   Figure 1: Information syste m in the realworld context  [29]  ![]() 92                             Mary Levis, Markus Helfert and Malcolm Brady  Copyright © 2008 SciRes                                                                                 JSSM  Wang and Strong proposed a data quality framework  that includes the categories of intrinsic data quality, ac- cessibility data quality, contextual data quality and repre- sentational data quality outlined in table 1.  Table1. IQ dimensions [17]  DQ Category DQ Dimensions  Intrinsic DQ  Accuracy, Objectivity,  Believability, Reputation  Accessibility DQ Accessibility,  Access Security  Contextual DQ  R e le vancy , Value Added  Timeliness, Completeness,  Amount of Data  Representational  DQ  Interpretability, Ease of  understanding, Concise  Representation,  Consistent Representation  The quality of websites may be linked to such criteria  as timeliness, ease of navigation, ease of access and  presentation of information. From the customer's perspec- tive usability is the most important quality of a Web ap- plication [8].  However, even if all procedures are adhered to, errors  can still arise that reduce the quality standard of the online  experience . Fo r e xam pl e, a  fil e  m ay  be moved  or a n  image  deleted, which results in broken links. The root cause that  leads to web application  problems is the poor appro ach to  web design.   To remedy this several techniques exist to evaluate the  quality of websites for exam ple link checkers, accessi bility  checkers and cod e validation. To  help improve th e quality  of a website, aspects such as structure and page layout  need to be consistent and coherent. A good website must  include safe guards a gainst fa ilure an d prov ide si mple , user  friendly data entry and validation processes.   From the literature reviewed a universal definition of  information quality is difficult to achieve [3, 21, 26, 29, 38,  42]. Accor ding to [25]  'Technically, information that meets  all the requirements is quality information'.  Some ac- cepted definitions of quality from the quality gurus are  shown in table 2.   Table 2. Quality definitions from the quality gurus  Author Quality Definitions  Deming Meeting the customers needs  Juran Fitness for use  Crosby Conformance to requirements  Ishikawa Continuous improvement  Feigenbaum Customer satisfaction  One definition of quality is ‘the totality of characteris- tics of an entity that bear on its ability to satisfy stated and  implied needs’ [13, 14]. Two requirements for website  evaluation emerge from this definition. 1) general valua- tion of all the site’s characteristics and 2) how well the site  meets specific needs.  1.1. Related Work  Pernici and Scannapieco [28] d iscuss a set of data quality  dimensions such as expiration, completeness, source reli- ability, and accuracy to evaluate data quality in web in- formation systems to support correct interpretation of web  pages content. Cusimano Com Corp [4] declared that ef- fective web sites must be clear, informative, concise, and  graphically appealing.  Tilton [40] recommends that Web designers should  present clear information that has a consistent navigation  structure. Hylnka and Welsh [12] put forward the argu- ment that the web page is a sour ce of communication and  should be analysed within communication theory.  Kelly and Vidgen [16] is concerned with the combina- tion of a quality assessment method, E-Qual, and a light- weight quality assurance method, QA focus and states that  website developers need to use standards and best prac- tices to ensure that websites are functional, accessible and  interoperable.  There are a number of ways to evaluate the quality of  websites, such as competitive analysis, inspection, and  online questi onnaires.  WebQu al, develope d by Ba rnes and  Vidgen [1] is one approach to the assessment of website  quality. WebQual, has 3 main dimensions: usability, in- formation quality, and service interaction quality. Ac- cording to [1], WebQual is a ‘structured and disciplined  process that provides a means to identify and carry the  voice of the c ustomer thr ough each stage  of product a nd or  service development and implementation’.  Usability is concerned with the quality associated with  the site design; Information Quality is con cerned with the  quality of content of the site; Service In teraction  quality is  concerned with quality prov ided to the users as they enter  into the site. Within these dimensions, WebQual consists  of a set of 23 questions regarding the website being as- sessed and each question contains a rating from 1-7; 1 =  strongly disagree, 7 = strongly agree.  Detailed information about evaluating websites can be  found at [24, 34, 35]. Eppler & Muenzem ayer [7] identifies  2 manifestations, 4 quality categories, and 16 quality di- mensions. Kahn et al. [17] mapped IQ dimensions to the  PSP/IQ model with 2 quality types, 4 quality classifica- tions, and 16 quality dimensions. Zhu & Gauch [44] out- lines 6 quality metrics for information retrieval on the web.  1.2. What Is a Quality Website?  Online interactivity is a valuable way of improving the  ![]() Website Design Quality and Form Input Validation:                          93  An Empirical Study on Irish Corporate Websites  Copyright © 2008 SciRes                                                                                 JSSM  quality of b usiness websites and web designers shou ld be  aware of how design affects the quality of the website and  the image of the organization.  Good websites have a rich and intuitive link structure. A  link going to the Customer Service should be named  'Customer Service' and the surfer looking for Customer  Service information will know this link goes to the page  they want. T herefore, ' click he re' should ne ver be use d as a  link.  Information managers and developers must determine  how much information users need [25]. Some users will  need much background on a specific topic whilst others  may only need a summary or overview. A good web de- signer will think clearly about how each  piece of data links   up with the rest of the content on the website and will  organize the links accordingly.  Without a clear navigation system, viewers can become  disoriented.  Hyperl inks  are di stinguis hed fr om  norm al text  within a page by its colour. When the page pointed to by a  hyperlink has been 'visited’ browsers will inform the users  by changing the link's colour [41].  The most vulnerable part of any web application is its  forms and the most common activity of web applications is  to validate the users’ data. According to the Open Web  Applications Security Project [27] invalidated input is in  the top ten critical web application security vulnerabilities.  Input validation is an important part of creating a robust  technological system and securing web applications. Be- cause of the fundamental client server nature of the web  application, input validation should be done both on the  client and the server.  Client side validation is used to provide input data  validation at the data collection point before the form is  submitted and check that the required fields are filled and  conform to certain characteristics such as built in length  restriction, numeric limits, email address valid ity, charac- ter data format etc. Incorrect data validation can lead to  data corruption.  Table 3. Example validation checks  Validation check Description  Character set Ensure data only contain char- acters you expect  Data format Ensure structure of data is con- sistent with what is expected  Range check Data lies within specific range of  values  Presence check No missing / empty fields  Consistency check If title is 'Mr' then gender is  'Male'  Input validation should be performed on all incoming  data ensuring the information system stores clean, correct  and useful data. Examples of invalid data are: text entered  into a numeric field, numeric data entered into a text field,  or a percentage entered into a currency field.  Table 3  provides an example set of  checks that coul d be perform ed  to ensure the incoming data is valid before data is proc- essed or used .  Having contact information available and visible on the  website is a marketing plus that potential customers use in  order to judge a company’s trustworthiness, as it signifies  respect for the customer and implies promise of good  service.  Feedback mechanisms built into the website are a useful  way to get meaningful feedback on the website and service  quality from the people who matter most – your customers.  After all, one definition of quality is ‘meeting or exceeding  the customer’s expectations’.  One of the most important factors for a website being  successful is speed. If the website is unresponsive, with  long response times the visitors will not come again. Speed  or responsiveness is integral to good website design and  organizational success. Web pages should be designed  with speed in mind [31]. It is estimated that if a page  doesn't load within 5-8 seconds you will lose 1/3 of your  visitors [35]. However, many designers believe that with  the recent development of broadband, visual aesthetics is  now more im portant as downl oad speed is not  such a major  concern. Nevertheless, not all users have broadband and  this should be taken into cons ideration .  For the pu rpos e of t his st udy   we c ond ucted  an em pi rical   study using a data set of twenty one finalists in a recent  website quality technology award. The aim of this study  was to examine these  websites for Technical quality issues  from the user’s perspective.  Our analysis focused on helping website owners under- stand the importance of certain website characteristics,  quality of information and functionalities. During the  analysis we tested functionalities in the website like forms,  the navigation process, the relevance of all click through  and the page download speed.  The rest of the paper is organized as follows: Sectio n 2  shows our methodology, Section 3 giv es a brief summary  and Section 4 some conclusions.  2. Research   2.1. Methodology  We conducted an empirical study on a recent accountancy  website quality technology award competition using the  full data set of twenty one finalists that included (3) Char- ity/Not for Profit organizations, (7) Large Quoted Com- panies, (2)  Sma ll Quot ed C om panies an d (9)  Statut ory an d  Unquoted Companies. The identity of websites has been  concealed due to confidentiality regulations regarding  ![]() 94                             Mary Levis, Markus Helfert and Malcolm Brady  Copyright © 2008 SciRes                                                                                 JSSM  38% 10% 5% 29% 52% 29% 90% 48% 57% 62% 90% 95% 71% 48% 71% 10% 52% 43% 0% 10% 20% 30% 40% 50% 60% 70% 80% 90% 100% their identity.  The aim of this study was to examine these websites for  Technical quality issues. This required validatin g the sites  against a series of checkpoints that included:  checking  that legal and regulatory guidelines were adhered to (e.g.  data protection and privacy), that pages conformed to  Web-Accessibility standard (e.g. missing 'alt tags'), miss- ing page titles, browser compatibility, user feedback  mechanisms, applications were functioning correctly (e.g.  online forms are validated for input etc.). It also included  evaluation of the main characteristics and structure of the  sites for example clear ordering of information, broken  links, and ease of navi gation. The principle us ed was based  on the same criteria used t o evaluate the participants in the  2006 award [10, 32, 39, 43]. The criteria subset used for  this study is outlined in tab le 4.   Table 4. Set of criteria used in our study  Validation Criteria  Contrast colours support readability & understanding Professional appearance  Do not use 'click here'  What you clicked on is title of page jumped to  Links back to home page are functional & relevant  Help features available and easy to access  Visited links change colour  Site map  Interactive form validated for input  Mailto parameters set correctly?  Web address simply a case of adding .com or .ie to  Useful search   eng ine provi ded  Site search provided  FAQ  Data Protection & Privacy  2.2. Findings and Analysis  Table 5 shows the number of companies who defaulted  and the number of companies who adhered to selected  criteria.  Twelve websites did not include a link to their data  protection and privacy policy. A help and Frequently  Asked Question (FAQ) page is a general requirement for  good website design. As far as navigation goes, this page  should tell the user how to find products or information  and how to get to the sitemap, yet, 10 companies did not  have a FAQ link and 11 did not have help features avail- able and easy to access.   Seven out of the 21 sites evalu- ated did not have the mailto parameters set correctly.  Fifteen sites had fully functional and relevant links to  other pages and back to the homepage. Thirteen sites  promoted contrast colours supporting readability and un- derstanding and 19 had a professional feel and appearance  and did not  ha ve  horizo ntal  s croll  bars . T went y o f the  tot al  twenty one sites adhered to the criteria of having the title of  page jumped also as the labe l of the link connecting to it.  The percentage of sites that adherence to the criteria and  the percentage of sites that defaulted on the criteria are  shown in figure 2.  Table 5. Criteria for website evaluation  Validation Criteria De- faulted Ad- hered Contrast colours support read- ability and understanding 8 13  Professional appearance 2 19  No use of 'click here' links 12 9  What is clicked on is title of  page jumped to 1 20  Links to home page functional  and relevant 6 15  Help features available & easy to  access 11 10  Visited links change colour 16 5  Site map available 6 15  Form validation for input 17 4  Mailto parameters set correctly 7 14  Web address is a case of add- ing .com or .ie to company name 4 17  Useful search engine pr o vi de d 19 2  Site search provided 7 14  Frequently Asked Questions 10 11  Data Protection and Privacy 12 9  a   b    c   d    e   f    g    h   i  Figure 2. Percentage of sites that defaulted and per- centage of sites that adhered to selected criteria (a-  Colours s support readability; b-Professional appear- ance; c-what clicked on was title of page jumped to;  d-Links functional &relevant; e-Help features avail- able &easy to access; f-Mailto parameters set cor- rectly. g-Use search engine; h-Frequently asked ques- tion page; i-Data protection &privacy)  Figure 3 depicts the results from the Friendly URLs  ![]() Website Design Quality and Form Input Validation:                          95  An Empirical Study on Irish Corporate Websites  Copyright © 2008 SciRes                                                                                 JSSM  (Uniform Resource Locator). Seventeen (81%) of the 21  sites tested had good structured semantic URLs, made up  of the actual name of t he specific company where we  could  guess the UR L by si mply addi ng .com  ,  .org or .ie  to the  company name.  For example a company named ‘Jitnu’  had a URL http://www.jitnu.ie or http://www.jitnu.co m or  http://www.jitnu.org as the web addresses which convey  meaning and structure. Only 19 % of the companies ex- amined defaulted on these criteria having a URL for ex- ample such as http://www.jitnu.ie/?id=478 instead of  http://www.jitnu.ie/services  or had a file extension  like .php as part of their URL.   Web address adding .com, .org or .ie  Adhered to , 17, 81% Defaulted on, 4, 19% Figure 3. Res u l ts  o f  th e friendl y  URL's crite ria  Figure 4 shows that 16 (76%) of the 21 sites examined  used the same link colour for visited and unvisited pages  and did not support a convention that users expect. Failing  on this navigational aid could well increase navigational  confusion and introduce usability problems for the user.  Good practice is to let viewers see their navigation path  history (i. e. pages they have al read y  visited) by dis pl a y ing  links to ‘visi t ed pages’ in a different colour.  Visited links change colour Did not change  colour,   16, 76% Changed Co lo ur , 5, 24% Figure 4. Visited pages changed link colour  Sitemaps are particularly beneficial when users cannot  access all areas of a we bsite through the browsing interface.  Failure to provide this access option may lose potential  viewers. A large website should contain a site map and  search optio n. From analysis of  our findings in figu re 5, we  show that six websites ( 29% ) di d n ot  provide a site map.  Visitors appreciate search capability on sites that deals  with several different products or services. In figure 6 we  show that although adding a search function on a website  helps visitors to quickly find information they need, seven  (33%) of the 21 sites reviewed failed to provide a com- prehensive site search or search interface.   Defaulted,  Site map , 6 Adheared to,  Site map , 15 0 2 4 6 8 10 12 14 16 Site map  Figure 5. Site map  While creating a good navigation system will be suffi- cient help for many people, it won't meet the needs of  everyone . It ap pears that  these  companies fail to realise t he  importance of providing a search capability, which not  only make sites more interactive but also gives visitors  more control over their browsing experience.   Provision of search facility No Search facility  provided, 19 Search Facility  provided,  2 0 2 4 6 8 10 12 14 16 18 20 Is a useful search engine provided Figure 6. Search facility  Figure 7 shows the results of checking user-entered  email addresses for valid inpu t.  An email address should  contain one and only one (@) and also c o ntain at  l e ast one  (.).    There should be no spaces or extra (@). There must  be at least one (.) after the (@) for an email address to be  valid. Some websites had implemented some form of  email address validation bu t did so incorrectly. Fo r exam- ple they correctly rejected jitnu.eircon.net and  jitnu@eircom@net as invalid email addresses, however,  they incorrectly accepted 'jitnu.eircom@net', as a valid  email address thus allowing an invalid email address to  pass to the system as a valid. While they co rrectly checked  for the presence of the (@) and the (.), they did not how- ever check the order in which the (@) and the (.) appeared  ![]() 96                             Mary Levis, Markus Helfert and Malcolm Brady  Copyright © 2008 SciRes                                                                                 JSSM  in the inputted email addr ess. From the review of the sites  we found that 17 (81%) had no validation process on email  addresses while only 4, (19%) of the 21 sites reviewed  shown in figure 7 had complete validation.  Interactive form validated for input  Passed input  validation,  4, 19% Failed input  validation ,  17, 81% Figure 7.  Email validation  Figure 8 shows that 12 (57%) of the 21 websites were  careless about their link text quality by using the ‘Click  Here’, which does not  give in dicat ion of the conte nt of t he  linked page, while nine (43%) used meaningful link text  which clearly identified the target of the links.  Defaulted:  Used  'click here',  12 Adheared to,   Did not use 'click here',  9 0 2 4 6 8 10 12 14  Use typed link labels on HTML anchors; do not use 'click here' Figure 8. Click here anchors  Figure 9 shows the number of companies that observed  and adhered to the quality criteria set out for this review  and figure 10 shows the number that defaulted in the above  criteria. It can be seen that 19 sites had a professional  appearance wi th no annoying horizont a l  scroll bars  and 20  sites used the page title of th e page link ed to as an an chor.  However only 2  of th e 21  sites un d er r eview h ad pr ov ided  a site search option and 4 sites had complete validation  on  email addresses.  Figure 11 shows that six out of 21 sites under review did  not have their ‘mailto’ parameter set correctly to facilitate  the user with easy feedback option and none of the 21 sites  provided an option for the users to view the feedback  provided by other users. Providing feedback options  greatly enhances the website as many errors go unnoticed  until the user draws it to the attention of the organization .  From our analysis we found that all sites had good page  load speed between a minimum 0.5 seconds and a maxi- mum of 1.5 seconds and an average load speed of .89  seconds. The average download per Kb was .26 seconds  with a minimum speed of .02 seconds and a maximum  speed of 3.76 seconds. The mini mum si te si ze wa s .17k b,  the maximum site size was 45.68kb and the average site  size was 20.45 kb.  Adhered To Criteria 13 19 20 15 10 15 2 11 9 0 5 10 15 20 25    a    b   c    d   e   f    g   h    i  Figure 9. Adhered to selected criteria(a- Colors sup- port readability; b-Professional appearance; c-what  clicked on was title of page jumped to; d-Links func- tional &relevant; e-Help features available &easy to  access; f-Mailto parameters set correctly. g-Use search  engine; h-Frequently asked question page; i-Data  protection &privacy)  Defaulted 8 21 6 11 6 19 10 12 0 2 4 6 8 10 12 14 16 18 20         a   b   c    d   e   f    g   h   i  Figure 10. Defaulted on selected criteria(a- Colors  support readability; b-Professional appearance;c-what  clicked on was title of page jumped to; d-Links func- tional &relevant; e-Help features available &easy to  access; f-Mailto parameters set correctly. g-Use search  engine; h-Frequently asked question page; i-Data  protection &privacy)  3. Summary  The Internet is an interactive channel and successful web- sites are those that are built on a foundation of interactiv- ity. As the Internet's first real feedback mechanism, forms  processing is still the most widely used interactivity on  websites.  Companies have invested vast amounts of  ![]() Website Design Quality and Form Input Validation:                          97  An Empirical Study on Irish Corporate Websites  Copyright © 2008 SciRes                                                                                 JSSM  money and resources to make their website a strategic  part of their business. But what do their customers actu- ally experience when they perform a transaction on their  site?  From our study we found that th e web applications un- der investigation were notorious for taking practically any  type of input, assuming that it's valid, and processing it  further. Not validating input is one of the biggest mistakes  that Web-application developers make. This can lead to  database corruption. The results from our investigation of  the company websites show that one of the biggest prob- lems with online forms on the web is that many corporate  companies failed to validate user input to ensure reliability  and potentially resulted in a database full of useless in- formation. (See figure 12, figure 13 and figure 14).  Number/21 Number/21, 6 Number/21,  21 0 5 10 15 20 25   Feedback " mailto" parameters not set correctly users can not see the feedback that  ot he rs  have provided    Feedback " mailto" parameters not set  c orrec t ly  users c an not see t he feedback that  others  have provided  Figure 11. Feedback option and accessibility  Figure: 12. Example of no date of birth validation on  corporate website form  Figure 13. Example of trusting the user to enter and  validate their own data  Figure 14. Example of useless information collected  in one Irish corporate website  This study highlighted that to date a large number of  web applications have not used mechanisms to validate  even the very basic data input at the source of collection.    Given that the sites under review in this study included  large quoted companies, small quoted companies, charities  and not for profit, statutory and unquoted organizations  and that some had been recognized for excellence in fi- nancial reporting; it was surprising to find that 81% of the  sites under examination failed on basic input validation.  All 100% of large quoted companies and 100% small  quoted companies failed in their email input validation  while 67% of charities/not for profit organizations and  67% of statutory and unq uo ted or gan izations und er inv es- tigation failure to valid ate emails. No less than  90 % failed  to provide a useful search engine but 71 % did provide a  ![]() 98                             Mary Levis, Markus Helfert and Malcolm Brady  Copyright © 2008 SciRes                                                                                 JSSM  site map.  Providing a site search function makes the site search- able. The sitemap should include every page on the site,  categorized for easier navigation. These are the links that  users look for  when they cannot  find what they are actually  looking for on the site. However, 67% provided a site  search facility and 81% had friendly URL’s that were easy  to remember and most sites had a good de sign layout that  was consistent throughout. The consistency aspect of  quality was closely adhered to by all sites making it easier  for the user to navigate.  4. Conclusions  Today's Internet user expects to experience personalized  interaction with websites. If the company fails to deliver  they run the risk of losing a potential customer forever.  An important aspect of creating interactive web forms to  collect information from users is to be able to check that  the information entered is valid, therefore; information  submitted through these forms should be extensively  validated. Validation could be performed using client  script where errors are detected when the form is submit- ted to the server and if any errors are found the submis- sion of the form to the server is cancelled and all errors  displayed to the user. This allows the user to correct their  input before re-submitting the form to the server. We can  not underestimate the importance of input validation  which ensures that the application is robust against all  forms of input data obtained from the user.  Although the majority of web vulnerabilities are easy to  understand and avoid many web developers are unfortu- nately not very security-aware. A company database needs  to be of reliable quality in order to be usable. A simple  check whether a website conforms to the very basic stan- dards could have been done using the W3C HTML vali- dation service, which is free to use.  Web developers need to become aware and trained in  Information Quality Management principles, and espe- cially in the information quality dimensions as outlined in  Table 2. The only proven reliable way to dea l with bad data  is to prevent it from entering the system. Input can be  compared against a specific value; ensure that an input  field was filled and that the value falls within a certain  range. Allowing bad data into the system makes the en tire  system unreliable and indeed unusable.  Making purchases online is all about confidence; a  customer must feel assured that you are a reputable com- pany, and the best way to project that image is through a  well designed website. A consumer visiting a website that  looks a little dodgy will not feel confident enough to  submit their credit card information.  Slow response times and difficult navigation are the  most common complaints of Internet users. After waiting  past a certain ‘attention threshold’ users look for a faster  site. Of course, exactly where that threshold is d epends on  many factors. How compelling is the experience? Is there  effective feedback? Etc. Our analysis identified these and  many other shortcomings that should have been realised  and dealt with during the website test phases. Many  problems could be eliminated by checking for letters (al- phabet entries only); checking for numbers (numeric en- tries only); checking for a valid range of values ; checking  for a valid date input; and checking for valid email ad- dresses. However, it is important to keep in mind that a  user could enter a valid e-mail address that does not actu- ally exist. It is therefore, imperative that some sort of ac- tivation process needs to be done in order to confirm a  valid and correct email add ress.  5. Acknowledgements  This research is funded by Dublin City University Busi- ness School.  REFERENCES  [1] Barnes S. J., and Vidgen R.T., An Integrative Ap- proach to the Assessment of E-Commerce Quality,  Journal of Electronic Commerce Research, Vol. 3,  No. 3, (2002) pp.114-27.  [2] Beckford J., 2nd edition, Quality, Rutledge Taylor  and Frances Group, London and New York (2005).  [3] Bugajski J., Grossman R. L., Tang Zhao., An event  based framework for improving information quality  that integrates baseline models, casual models and  formal models, IQIS 2005 ACM 1-59593-160-0/5/06.  (2005).  [4] Cusimano Com Corp., Planning and Designing an  Effective Web Site,  http://www.cusimano.com/webdesign/articles/effecti v.html .  [5] Clarke R., Internet Privacy Concerns Confirm the  Case for Intervention, Communications of the ACM,  Vol. 42, No. 2, (1999) pp.60-68.  [6] Cranor L.F., Internet Privacy, Communications of  the ACM, Vol. 42, No. 2, (19 9 9) , pp.28-31.  [7] Eppler, M & Muenzenmayer, P, Measuring informa- tion quality in the web context: A survey of state of  the art instruments and an application methodology.  Proceedings of 7th international conference on in- formation quality, (2002), pp.187-196.  [8] Fraternali, P., Tools and Approaches for Developing  Data-Intensive Web Applications: A Survey, ACM  Computing Surveys, Vol.31, No.3, (1999).   [9] Internet World Statistics  http://www.internetworldstats.com.  [10] FFHS Web Award  http://www.ffhs.org.uk/awards/web/criteria.php .  [11] Gefen D., E-Commerce: The Role of Familiarity and  Trust, Omega, Vol. 28, No. 6, (2000), pp.725-737.  [12] Hlynka D., and Welsh J., What Makes an Effective  ![]() Website Design Quality and Form Input Validation:                          99  An Empirical Study on Irish Corporate Websites  Copyright © 2008 SciRes                                                                                 JSSM  Home Page? A Critical Analysis  http://www.umanitoba.ca/faculties/education/cmns/a ect.html).   [13] ISO 8402, Quality Management Assurance- Vo- cabulary, Int’l Org. for standardization, (1994)   http://www.techstreet.com/iso_resource_center/iso_ 9000_meet.tmpl .  [14] ISO 9241- 11 , Int ernational Or gani zati o n fo r St an- dardization, Geneva, (1998).  http://www.idemployee.id.tue.nl/g.w.m.rauterberg/le cturenotes/ISO9241part11.pdf .  [15] Jarvenpaa S.L., N. Tractinsky, and M. Vitale, Con- sumer Trust in an Internet Store, Information Tech- nology and Management, Vol. 1, No. 1, (2000),  pp.45-71.  [16] Kelly B., and Vidgen R., A Quality Framework for  Web Site Quality: User Satisfaction and Quality As- surance, WWW 2005, May 10-14, 2005, Chiba, Ja- pan, ACM 1-59593-051-5/05/0005, (2005).  [17] Khan B K ., Strong D.M. & Wang R.Y., Information  Quality benchmarks: Product and service perform- ance, Communications of the ACM, Vol. 45, No 4,  (2002), pp.84-192.  [18] Kumar G, Ballou T, Ballou D, P.,Guest editors,  Examining data Quality, Communications of the  ACM, Vol. 41, No 2, (1998), pp.54-57.  [19] Useit.com: http://www,useit.com/ .  [20] Manchala D.W., E-Commerce Trust Metrics and  Models, IEEE Internet Computing, Vol. 4, No. 2,  (2000), pp.36-44.  [21] Mandel T., Quality Technical Information: Paving  the Way for UsableW3C Web Content Accessibility  Guidelines 1.0,  http://www.w3.org/tr/wai-webcontent/ .  [22] McKnight, D.H., L.L. Cummings, and N.L. Cher- vany, Initial Trust Formation  in New Organizational  Relationships, Academy of Managemen t Review, Vol.  23, (1998), pp.473-490.  [23] McKnight D.H., and N.L. Chervany, An Integrative  Approach to the Assessment of E-Commerce Qual- ity, Journal of Electronic Commerce Research, Vol.  3, No. 3, (2002), pp.114-127.  [24] Web standard schecklist.  http://www.maxdesign.com.au/presentation/checklist .htm.  [25] Olson J.E., Data Quality: The Accuracy dimens ion,  Morgan Kaufmann, ISBN 1558608915, (2003).  [26] Open W eb Application Security Project,  http://umn.dl.sourceforge.net/sourceforge/owasp/O WASPTopTen2004.pdf .  [27] Orr K., Data Quality and Systems, Communications  of the ACM, Vol. 41, No 2, (1998), pp.66-71.  [28] Pernici B., and Scannapieco M., Data Quality in  Web Information Sys tems, (2002) 397-413.  [29] Pike R.J., Barnes R., TQM in Action: A practical  approach to continuous performance improvement,  Springer, ISBN 0412715309 (1996).  [30] Mandel T., Print and Web Interface Design, ACM   Journal of Computer Documentation, Vol. 26, No. 3.  ISSN: 1527-6805, (2002).  [31] Poll R., and te Boekhorst, P., Measuring quality  performance measurement in libraries, 2nd revised.,  Saur, Munchen, IFLA publications Durban, 127,  (2007) http://www.ifla.org/iv/ifla73/index.htm.  [32] Quality Criteria for Website Excellence World Best  Website Awards  http://www.worldbestwebsites.com/criteria.htm.  [33] Redmond T.C., Improve Data Quality for Competi- tive Advantage, Sloan Management Review, Vol. 36,  No 2, (1995), pp.99-107.  [34] Schulz U., Web usability,  http://www.bui.fh-hamburg.de/pers/ursula.schulz/we busability/webusability.html,   [35] Selfso  http://www.selfseo.com/website_speed_test.php.  [36] Strong D. M., Lee Y. W., Wang R,Y., Data Quality  in Context Communications of the ACM,  Vol. 40,  No 5, , (1997), pp.103-109.  [37] Stylianou A.C., Kumar R.L., An integrative frame- work for IS Quality management, Communications  of the ACM, Vol. 43, No 9, (2000), pp.99-104.  [38] Tauscher L., Greenberg S., How people revisit web  pages: Empirical findings and implication for the  design of history systems, International Journal of  Human-Computer Studies, Vol. 47, (1997),  pp.97-137.  [39] McClainT, http://www.e-bzzz.com/web_design.htm.  [40] Tilton J. E., Composing Good HTML,  http://www.cs.cmu.edu/~tilt/cg h/.  [41] Wang R.Y., and Strong D.M. Beyond accuracy:  what data quality means to data consumers, Journal  of Management Information Systems Vol. 12, No 4,  (1996), pp.5–34.  [42] Wang R.Y., A product perspective on Total Data  Quality Management, Communications of the ACM,  Vol.41, No.2, (1998), pp.58-65.  [43] Web standards  http://www.webstandard sawards.com/criteria/#Crite ria .  [44] Zhu X., and Gauch S., Incorporating quality metrics  in centralized distributed information retrieval on  the World Wide Web. Proceedings of the 23rd an- nual international ACM SIGIR conference on Re- search and development in information retrieval,  Athens, Greece. (2000), pp.288-295.  ![]() 100                             Mary Levis, Markus Helfert and Malcolm Brady  Copyright © 2008 SciRes                                                                                 JSSM  AUTHORS’ BIOGRAP HIES  Mary Levis joined Dublin City University Business School as a full time, Ph.D Research Scholar in the field of Info r- mation Management, in November 2006. Mary’s work is co-supervised by Dr. Malcolm Brady, Dublin City University  Business School, and Dr. Markus Helfert, Dublin City University (School of Computing). Her studies are funded for  three years by a scholarship from Dublin City University Business School. She holds a B.Sc degree in Computer Ap- plications (Info Sys) from Dublin City University (School of Computing). Mary has presented her work at many Inter- national conferences. Mary is a member of the Irish Computer Society (ICS), British Computer Society (BCS), the In- ternational Association of Information Data Quality (IAIDQ), Engineers Ireland (EI) and United Kingdom Academy for  Information Systems (UKAIS) Institute for the Management of Information Systems. Mary has been invited by the In- ternational Conference on Enterprise Information Systems (ICEIS) to be session chair of an ICEIS-2008 session(s), in  Barcelona, Spain. Email: Mary.levis2@mail.dcu.ie  Dr. Malcolm Brad y  has l ectured in manag ement at Dubli n C i ty  Uni versity B usiness School since 19 9 6. P ri o r t o joining  DCU he worked for m any years as an IT consultant in the fi nancial services and utility indust ries and before that as a ci vil  engineer i n the desig n and const ruction of g as, water, a nd sewera ge systems. He teaches c ourses in strat egic ma nagement,  project management, and business process innovation. His research interests are in game theory, duopoly, advertising,  differentiation and competitive advantage. He graduated from University College Dublin with a Bachelors degree in Civil  Engineering and a Masters degree in Management Science; he holds a Diplôme d'Ingénieur from Ecole National  Supérieur du Pétrole et des Moteurs, a French Grande Ecole; he obtained an MBA from Dublin City University. He  recently completed a PhD from the Department of Economics at Lancaster University. Email: malcolm.brady@dcu.ie  Dr. Markus Helfert is a lecturer in Information Systems at Dublin City University, Irelan d and  programme ch air of  the  European M.Sc. in Business Informatics at Dublin City University. His research interests include Information Quality,  Data Warehousing, Information System Architectures, Supply Chain Management and Business Informatics education.  His current research in information quality builds on his PhD research in data quality management in data warehouse  systems. He holds a Doctor in business administration from the University of St. Gallen (Switzerland ),  a Master-D egr ee  in business informatics from the University Mannheim (Germany) and a Bachelor of Science from Napier University,  Edinburgh (UK-Scotland). He has authored academic articles and book contributions and has presented his work at in- ternational conferences. Email: markus.helfert@dcu.ie   | 
	











