In this communication a design of embedded web server based on the Ethernet technology for remote monitoring of weather parameters is presented. This web server monitors parameters viz. temperature and humidity and transmits this... more
In this communication a design of embedded web server based on the Ethernet technology for remote monitoring of weather parameters is presented. This web server monitors parameters viz. temperature and humidity and transmits this information in the form of HTML web-page. The input sensors LM35 semiconductor temperature sensor and SY-HS-220 humidity module have been employed, providing accuracy about 1° centigrade and 2% relative humidity. The web server provides simultaneous access to multiple nodes on the network.
In this paper, we propose a hybrid approach of Arabic scripts web page language identification based on decision tree and ARTMAP approaches. We use the decision tree approach to find the general identities of a web document, be it an... more
In this paper, we propose a hybrid approach of Arabic scripts web page language identification based on decision tree and ARTMAP approaches. We use the decision tree approach to find the general identities of a web document, be it an Arabic script-based or a non-Arabic-based. Then, we use the selected representations of identified pages from the decision tree approach as
Web archives preserve the history of born-digital content and oer great potential for sociologists, business analysts, and legal experts on intellectual property and compliance issues. Data quality is crucial for these purposes. Ideally,... more
Web archives preserve the history of born-digital content and oer great potential for sociologists, business analysts, and legal experts on intellectual property and compliance issues. Data quality is crucial for these purposes. Ideally, crawlers should gather sharp captures of entire Web sites, but the politeness etiquette and completeness requirement mandate very slow, long-duration crawling while Web sites undergo changes. This
The internet has become one of the most important means of communication in all areas of our life. In the paper we focused on central and local government bodies and their attitude towards information and communication technology. By... more
The internet has become one of the most important means of communication in all areas of our life. In the paper we focused on central and local government bodies and their attitude towards information and communication technology. By analysing web pages, inquiring public servants and testing the responses on citizens' questions we tried to discover influences of Internet on better informing of citizens, their participation in making decisions of public interest and communication between citizens and central and local government bodies.
A new multidisciplinary design and entrepreneurship course is being offered at Bilkent University. The course entitled "Innovative Product Design and Development I-II" is open to senior students of six different departments. It... more
A new multidisciplinary design and entrepreneurship course is being offered at Bilkent University. The course entitled "Innovative Product Design and Development I-II" is open to senior students of six different departments. It is a two-semester course that brings together students from Departments of Computer Engineering, Economics, Electrical Engineering, Graphic Design, Industrial Engineering, and Management. A team of six professors, one
We describe Syskill & Webert, a software agent that learnsto rate pages on the World Wide Web (WWW), decidingwhat pages might interest a user. The user rates exploredpages on a three point scale, and Syskill & Webert learnsa user profile... more
We describe Syskill & Webert, a software agent that learnsto rate pages on the World Wide Web (WWW), decidingwhat pages might interest a user. The user rates exploredpages on a three point scale, and Syskill & Webert learnsa user profile by analyzing the information on each page.The user profile can be used in two ways. First, it can beused to
Modern web search engines are expected to return the top-k results efficiently. Although many dynamic index pruning strategies have been proposed for efficient top-k computation, most of them are prone to ignoring some especially... more
Modern web search engines are expected to return the top-k results efficiently. Although many dynamic index pruning strategies have been proposed for efficient top-k computation, most of them are prone to ignoring some especially important factors in ranking functions, such as term-proximity (the distance relationship between query terms in a document). In our recent work [Zhu, M., Shi, S., Li,
In this paper we present an English grammar and style checker for non-native English speakers. The main characteristic of this checker is the use of an Internet search engine. As the number of web pages written in English is immense, the... more
In this paper we present an English grammar and style checker for non-native English speakers. The main characteristic of this checker is the use of an Internet search engine. As the number of web pages written in English is immense, the system hypothesizes that a piece of text not found on the Web is probably badly written. The system also hypothesizes that the Web will provide examples of how the content of the text segment can be expressed in a gramatical and idiomatic way. So, after the checker warns the user about the odd character of a text segment, the Internet engine searches for contexts that will be helpful for the user to decide whether he/she corrects the segment or not. By means of a search engine, the checker also suggests the writer to use expressions which are more frequent on theWeb other than the expression he/she actually wrote. Although the system is currently being developed for teachers of the Open University of Catalonia, the checker can also be useful for sec...
This paper introduces the concept of a classification tool for Web pages called WebClassify, which uses modified naive Bayesian algorithm with multinomial model to classify pages into various categories. The tool starts the classification... more
This paper introduces the concept of a classification tool for Web pages called WebClassify, which uses modified naive Bayesian algorithm with multinomial model to classify pages into various categories. The tool starts the classification from downloading training Web text from Internet, preparing the hypertext for mining, and then storing Web data in a local database. The paper also gives an
Today, many organizations maintain a variety of systems and databases in a complex ad-hoc architecture that does not seem to fulfill the needs for company-wide unstructured information management in business processes, business functions,... more
Today, many organizations maintain a variety of systems and databases in a complex ad-hoc architecture that does not seem to fulfill the needs for company-wide unstructured information management in business processes, business functions, and the extended enterprise. We describe a framework to implement Enterprise Content Management (ECM) in order to address this problem. ECM refers to the technologies, tools, and methods used to capture, manage, store, preserve, and deliver content (e.g. documents, graphics, drawings, web pages) across an enterprise. The framework helps to select content objects that can be brought under ECM to create business value and guide the IT investments needed to realize ECM. The framework was tested in a large high tech organization.
This paper describes an attack concept termed Drive-by Pharming where an attacker sets up a web page that, when simply viewed by the victim (on a JavaScript-enabled browser), attempts to change the DNS server settings on the victim’s home... more
This paper describes an attack concept termed Drive-by Pharming where an attacker sets up a web page that, when simply viewed by the victim (on a JavaScript-enabled browser), attempts to change the DNS server settings on the victim’s home broadband router. As a result, future DNS queries are resolved by a DNS server of the attacker’s choice. The attacker can direct the victim’s Internet traffic and point the victim to the attacker’s own web sites regardless of what domain the victim thinks he is actually going to, potentially leading to the compromise of the victim’s credentials. The same attack methodology can be used to make other changes to the router, like replacing its firmware. Routers could then host malicious web pages or engage in click fraud. Since the attack is mounted through viewing a web page, it does not require the attacker to have any physical proximity to the victim nor does it require the explicit download of traditional malicious software. The attack works under the reasonable assumption that the victim has not changed the default management password on their broadband router.
Web is a vast data repository. By mining from this data efficiently, we can gain valuable knowledge. Unfortunately, in addition to useful content there are also many Web documents considered harmful (e.g. pornography, terrorism, illegal... more
Web is a vast data repository. By mining from this data efficiently, we can gain valuable knowledge. Unfortunately, in addition to useful content there are also many Web documents considered harmful (e.g. pornography, terrorism, illegal drugs). Web mining that includes three main areas – content, structure, and usage mining – may help us detect and eliminate these sites. In this paper, we concentrate on applications of Web content and Web structure mining. First, we introduce a system for detection of pornographic textual Web pages. We discuss its classification methods and depict its architecture. Second, we present analysis of relations among Czech academic computer science Web sites. We give an overview of ranking algorithms and determine importance of the sites we analyzed.
Web accessibility rules, i.e., the conditions to be met by Web sites in order to be considered accessible for all, can be (partially) checked automatically in many different ways. Many Web accessibility evaluators have been developed... more
Web accessibility rules, i.e., the conditions to be met by Web sites in order to be considered accessible for all, can be (partially) checked automatically in many different ways. Many Web accessibility evaluators have been developed during the last years. For applying the W3C guidelines, their programmers have to apply subjective criteria, thus leading to different interpretations of these guidelines. As a result, it is easy to obtain different evaluation results when different evaluation tools are applied to a common sample page. However, accessibility rules can be better expressed formally and declaratively in rules that assert conditions over the markup. We have found that XSLT can be used to represent templates addressing many accessibility rules involving the markup of Web pages. Even more, we have found that some specific conditions relaying in the prose of the XHTML specification not previously formalized in the XHTML grammar (the official DTD or XML Schemas) could also be formalized in XSLT rules as well. Thus, we have developed WAEX as a Web Accessibility Evaluator in a single XSLT file. Such XSLT file contains 70+ singular accessibility and XHTML-specific rules not previously addressed by the official DTDs or Schemas from W3C.
Researchers in the ontology-design field have developed the content for ontologies in many domain areas. Recently, ontologies have become increasingly common on the World- Wide Web where they provide semantics for annotations in Web... more
Researchers in the ontology-design field have developed the content for ontologies in many domain areas. Recently, ontologies have become increasingly common on the World- Wide Web where they provide semantics for annotations in Web pages. This distributed nature of ontology development has led to a large number of ontologies covering overlapping domains. In order for these ontologies to be reused,
Abstract: ASP .NET web applications typically employ server controls to provide dynamic web pages, and data-bound server controls to display and maintain database data. Most developers use default properties of ASP .NET server controls... more
Abstract: ASP .NET web applications typically employ server controls to provide dynamic web pages, and data-bound server controls to display and maintain database data. Most developers use default properties of ASP .NET server controls when developing web applications, which allows for rapid development of workable applications. However, creating a high-performance, multi-user, and scalable web application requires enhancement of server controls using custom-made code. In this empirical study we ...
A mobile user may voluntarily disconnect itself from the Web server to save its battery life and avoid high communication prices. To allow Web pages to be updated while the mobile user is disconnected from the Web server, updates can be... more
A mobile user may voluntarily disconnect itself from the Web server to save its battery life and avoid high communication prices. To allow Web pages to be updated while the mobile user is disconnected from the Web server, updates can be staged in the mobile unit and propagated back to the Web server upon reconnection. We analyze algorithms for supporting disconnected write operations and develop a performance model which helps identify the optimal length of the disconnection period under which the cost of update propagation is minimized. The analysis result is particularly applicable to Web applications which allow wireless mobile users to modify Web contents while on the move. We also show how the result can be applied to real-time Web applications such that the mobile user can determine the longest disconnection period such that it can still propagate updates to the server before the deadline with a minimum communication cost
We present a detailed statistical analysis of the characteristics of partial Web graphs obtained by sub-sampling a large collection of Web pages. We show that in general the macroscopic properties of the Web are better represented by a... more
We present a detailed statistical analysis of the characteristics of partial Web graphs obtained by sub-sampling a large collection of Web pages. We show that in general the macroscopic properties of the Web are better represented by a shallow exploration of a large number of sites than by a deep exploration of a limited set of sites. We also describe and quantify the bias induced by the different sampling strategies, and show that it can be significant even if the sample covers a large fraction of the collection. ... 1. INTRODUCTION The number of Web pages that ...
Background: Publication records and citation indices often are used to evaluate academic performance. For this reason, obtaining or computing them accurately is important. This can be difficult, largely due to a lack of complete knowledge... more
Background: Publication records and citation indices often are used to evaluate academic performance. For this reason, obtaining or computing them accurately is important. This can be difficult, largely due to a lack of complete knowledge of an individual’s publication list and/or lack of time available to manually obtain or construct the publication-citation record. While online publication search engines have somewhat addressed these problems, using raw search results can yield inaccurate estimates of publication-citation records and citation indices. Methodology: In this paper, we present a new, automated method that produces estimates of an individual’s publicationcitation record from an individual’s name and a set of domain-specific vocabulary that may occur in the individual’s publication titles. Because this vocabulary can be harvested directly from a research web page or online (partial) publication list, our method delivers an easy way to obtain estimates of a publication-c...
The aim of this research study was to evaluate the websites of Jordan's universities from the usability perspective. Two online automated tools, namely: html toolbox and web page analyze were used along with a questionnaire directed... more
The aim of this research study was to evaluate the websites of Jordan's universities from the usability perspective. Two online automated tools, namely: html toolbox and web page analyze were used along with a questionnaire directed towards users of these websites. Tools were used to measure the websites internal attributes which can not be perceived by users, such as html code error, download time, and size of html page. The questionnaire was developed and designed based on 23 usability criteria divided into 5 categories. Each category deals with one usability aspect. The results showed that the overall usability level of the studied Websites is acceptable. However, there are some weaknesses in some aspects of the design, interface, and performances. Suggestions are provided in the study to enhance the usability of these websites.
Malaysia is a land and structure deformation active country due to long term rainy season which is common phenomena in monsoon tropical region. To secure this hazard under control, Land Stability Monitoring System (LandSMS), a web-based... more
Malaysia is a land and structure deformation active country due to long term rainy season which is common phenomena in monsoon tropical region. To secure this hazard under control, Land Stability Monitoring System (LandSMS), a web-based deformation ...