International Science Index
International Journal of Information and Communication Engineering
Improved Multi-Channel Separation Algorithm for Satellite-Based Automatic Identification System Signals Based on Artificial Bee Colony and Adaptive Moment Estimation
The applications of satellite-based automatic identification system (S-AIS) pave the road for wide-range maritime traffic monitoring and management. But the coverage of satellite’s view includes multiple AIS self-organizing networks, which leads to the collision of AIS signals from different cells. The contribution of this work is to propose an improved multi-channel blind source separation algorithm based on Artificial Bee Colony (ABC) and advanced stochastic optimization to perform separation of the mixed AIS signals. The proposed approach adopts modified ABC algorithm to get an optimized initial separating matrix, which can expedite the initialization bias correction, and utilizes the Adaptive Moment Estimation (Adam) to update the separating matrix by adjusting the learning rate for each parameter dynamically. Simulation results show that the algorithm can speed up convergence and lead to better performance in separation accuracy.
Derivation of an Approximate Location Estimate in Angle-of-Arrival Based Localization in the Presence of Angle-of-Arrival Estimate Error and Sensor Location Error
We derive closed-form expression of an approximate location estimate in AOA-based localization. In this paper, it is assumed that there is estimation error in AOA estimation and that estimates of sensor location can be expressed as zero-mean Gaussian random variable. The derivation is based on Taylor series expansion and an approximate solution of an over-determined linear system. The effect of sensor location error and AOA estimation error on the location estimate is explicitly shown in the derived expression, which implies that the result can be used for the closed-form expression of the mean-square-error of the location estimate.
Detectability Analysis of Typical Aerial Targets from Space-Based Platforms
In order to achieve effective detection of aerial targets over long distances from space-based platforms, the mechanism of interaction between the radiation characteristics of the aerial targets and the complex scene environment including the sunlight conditions, underlying surfaces and the atmosphere are analyzed. A large simulated database of the space-based radiance images is constructed considering several typical aerial targets, target working modes (flight velocity and altitude), illumination and observation angles, background types (cloud, ocean, and urban areas) and sensor spectrums ranging from visible to thermal infrared. The target detectability is characterized by the signal-to-clutter ratio (SCR) extracted from the images. The influence laws of the target detectability are discussed under different detection bands and instantaneous fields of view (IFOV). Furthermore, the optimal center wavelengths and widths of the detection bands are suggested, and the minimum IFOV requirements are proposed. The research can provide theoretical support and scientific guidance for the design of space-based detection systems and on-board information processing algorithms.
Electronic Payment Recording with Payment History Retrieval Module: A System Software
The Electronic Payment Recording with Payment History Retrieval Module is developed intendedly for the College of Science and Technology. This system software innovates the manual process of recording the payments done in the department through the development of electronic payment recording system software shifting from the slow and time-consuming procedure to quick yet reliable and accurate way of recording payments because it immediately generates receipts for every transaction. As an added feature to its software process, generation of recorded payment report is integrated eliminating the manual reporting to a more easy and consolidated report. As an added feature to the system, all recorded payments of the students can be retrieved immediately making the system transparent and reliable payment recording software. Viewing the whole process, the system software will shift from the manual process to an organized software technology because the information will be stored in a logically correct and normalized database. Further, the software will be developed using the modern programming language and implement strict programming methods to validate all users accessing the system, evaluate all data passed into the system and information retrieved to ensure data accuracy and reliability. In addition, the system will identify the user and limit its access privilege to establish boundaries of the specific access to information allowed for the store, modify, and update making the information secure against unauthorized data manipulation. As a result, the System software will eliminate the manual procedure and replace with an innovative modern information technology resulting to the improvement of the whole process of payment recording fast, secure, accurate and reliable software innovations.
Community Structure Detection in Networks Based on Bee Colony
In this paper, we propose a new method to find the community structure in networks. Our method is based on bee colony and the maximization of modularity to find the community structure. We use a bee colony algorithm to find the first community structure that has a good value of modularity. To improve the community structure, that was found, we merge communities until we get a community structure that has a high value of modularity. We provide a general framework for implementing our approach. We tested our method on computer-generated and real-world networks with a comparison to very known community detection methods. The obtained results show the effectiveness of our proposition.
Monitoring of Water Quality Using Wireless Sensor Network
Availability of portable water has been a global challenge especially to the developing continents/nations such as Africa/Nigeria. The World Health Organization WHO has produced the guideline for drinking water quality GDWQ which aims at ensuring water safety from source to consumer. Portable water parameters test include physical (colour, odour, temperature, turbidity), chemical (PH, dissolved solids) biological (algae, plytoplankton). This paper discusses the use of wireless sensor networks to monitor water quality using efficient and effective sensors that have the ability to sense, process and transmit sensed data. The integration of wireless sensor network to a portable sensing device offers the feasibility of sensing distribution capability, on site data measurements and remote sensing abilities. The current water quality tests that are performed in government water quality institutions in Benue State Nigeria are carried out in problematic locations that require taking manual water samples to the institution laboratory for examination, to automate the entire process based on wireless sensor network, a system was designed. The system consists of sensor node containing one PH sensor, one temperature sensor, a microcontroller, a zigbee radio and a base station composed by a zigbee radio and a PC. Due to the advancement of wireless sensor network technology, unexpected contamination events in water environments can be observed continuously. local area network (LAN) wireless local area network (WLAN) and internet web-based also commonly used as a gateway unit for data communication via local base computer using standard global system for mobile communication (GSM). The improvement made on this development show a water quality monitoring system and prospect for more robust and reliable system in the future.
Identifying the Risks on Philippines’ Pre- and Post-Disaster Media Communication on Natural Hazards
The Philippine is a hotbed of disasters and is a locus of natural hazards. With an average of 20 typhoons entering the Philippine Area of Responsibility (PAR) each year, seven to eight (7-8) of which makes landfall. The country rather inevitably suffers from climate-related calamities. With this vulnerability to natural hazards, the relevant hazard-related issues that come along with the potential threat and occurrence of a disaster oftentimes garners lesser media attention than when a disaster actually occurred. Post-disaster news and events flood the content of news networks primarily focusing on, but not limited to, the efforts of the national government in resolving post-disaster displacement, and all the more on the community leaders’ incompetence in disaster mitigation-- even though the University of the Philippines’ NOAH Center work hand in hand with different stakeholders for disaster mitigation communication efforts. Disaster risk communication is actually a perennial dilemma. There are so many efforts to reach the grassroots level but emergency and disaster preparedness messages inevitably fall short.. The Philippines is very vulnerable to hazards risk and disasters but social media posts and communication efforts mostly go unnoticed, if not argued upon. This study illustrates the outcomes of a research focusing on the print, broadcast, and social media’s role on disaster communication involving the natural catastrophic events that took place in the Philippines from 2009 to present. Considering the country’s state of development, this study looks on the rapid and reliable communication between the government, and the relief/rescue workers in the affected regions; and how the media portrays these efforts effectively. Learning from the disasters that have occurred in the Philippines over the past decade, effective communication can ensure that any efforts to prepare and respond to disasters can make a significant difference. It can potentially either break or save lives. Recognizing the role of communications is not only in improving the coordination of vital services for post disaster; organizations gave priority in reexamining disaster preparedness mechanisms through the Communication with Communities (CwC) programs. This study, however, looks at the CwC efforts of the Philippine media platforms. CwC, if properly utilized by the media, is an essential tool in ensuring accountability and transparency which require effective exchange of information between disasters and survivors and responders. However, in this study, it shows that the perennial dilemma of the Philippine media is that the Disaster Risk Reduction and Management (DRRM) efforts of the country lie in the clouded judgment of political aims. This kind of habit is a multiplier of the country’s risk and insecurity. Sometimes the efforts in urging the public to take action seem useless because the challenge lies on how to achieve social, economic, and political unity using the tri-media platform.
Students' Perceptions of Communication Design in Media: Case Study of Portuguese and Spanish Communication Students
The proliferation of mobile devices in society enables the media to disseminate information and knowledge more rapidly. Higher education students access these contents and share them with each other, in the most diverse platforms, allowing the ubiquity in access to information. This article presents the results and respective quantitative analysis of a survey applied to communication students of two higher education institutions: one in Portugal and another in Spain. The results show that, in this sample, higher education students regularly access news content believing traditional news sources to be more credible. Regarding online support, it was confirmed that free access to news is still the most popular mode. This study intends to promote the knowledge about the changes that occur in the relationship of higher education students with the media, characterizing the way in which the consumption of news by these students is processed, considering the resulting effects the evolution of digital media. It is intended to present not only the news sources they use, but also to know some of their habits and relationship with the news media.
Multi Tier Data Collection and Estimation, Utilizing Queue Model in Wireless Sensor Networks
In this paper, target parameter is estimated with desirable precision in hierarchical wireless sensor networks (WSN) while the proposed algorithm also tries to prolong network lifetime as much as possible, using efficient data collecting algorithm. Target parameter distribution function is considered unknown. Sensor nodes sense the environment and send the data to the base station called fusion center (FC) using hierarchical data collecting algorithm. FC builds underlying phenomena based on collected data. Considering the aggregation level, x, the goal is providing the essential infrastructure to find the best value for aggregation level in order to prolong network lifetime as much as possible, while desirable accuracy is guaranteed (required sample size is fully depended on desirable precision). First, the sample size calculation algorithm is discussed, second, the average queue length based on M/M[x]/1/K queue model is determined and it is used for energy consumption calculation. Nodes can decrease transmission cost by aggregating incoming data. Furthermore, the performance of the new algorithm is evaluated in terms of lifetime and estimation accuracy.
Internet Impulse Buying: A Study Based on Stimulus-Organism-Response Theory
As the advance of e-commerce technologies, the consumers buying behavior have changed. The focus on consumer buying behavior has already shifted from physical space to the cyberspace, which impulse buying is a major issue of concern. This study examines the stimulus effect of web environment on the consumer's emotional states, and in turn, affecting the urge of impulse buying based on a stimulus-organism-response (S-O-R) theory. Website ambiance and website service quality are the two stimulus variables. The study also explores the effects and the moderator effects of contextual variables and individual characteristic variables on the web environment, the emotional states and the urge of impulse buying. A total of 328 valid questionnaires were collected. Structural equation modeling was used to test the research hypothesis. This study found that both website ambiance and website service quality have a positive effect on consumer emotion, which in turn positively affect the urge of impulse buying. Consumer’s trait of impulse buying has a positive effect on the urge of impulse buying. Consumer’s hedonic motivation has a positive effect on both emotion state and the urge of impulse buying. On the other hand, the study found that money available for the consumer would positively affect consumer's emotion state and time available for the consumer would negatively affect the relationship between website service quality and consumer emotion. The result of this study validates Internet impulse buying behavior based on the S-O-R theory. This study also suggests that having a good website atmosphere and service quality is important to influencing consumers’ emotion and increasing the likelihood of consumer purchasing. The study could serve as a basis for the future research regarding online consumer behavior.
Self-Disclosure of Location: Influences of Personality Traits, Intrinsic Motivations and Extrinsic Motivations
With the popularity of smartphone usage and the flourish of social networks, many people began to use the 'check-in' functions to share their location information and days of live and self-disclosure. In order to increase exposure and awareness, some stores provide discounts and other benefits to attract consumers to 'check-in' in their stores. The purpose of this study was to investigate whether personality traits, intrinsic motivations, extrinsic motivations, and privacy concerns would affect self-disclosure of location for consumers. Research data were collected from 407 individuals that have used Facebook check-in in Taiwan. This study used SmartPLS 2.0 structural equation modeling to validate the model. The results show that information sharing, information storage, enjoyment, self-presentation, get a feedback, economic reward, and keep up with trends had significant positive effects on self-disclosure. While extroversion and openness to use have significant positive effects on self-disclosure, conscientiousness and privacy concerns have significant negative effects on self-disclosure. The results of the study provide academic and practical implications for the future growth of location-based self-disclosure.
Information Theoretic Approach for Beamforming in Wireless Communications
Beamforming is a signal processing technique extensively utilized in wireless communications and radars for desired signal intensification and interference signal minimization through spatial selectivity. In this paper, we present a method for calculation of optimal weight vectors for smart antenna array, to achieve a directive pattern during transmission and selective reception in interference prone environment. In proposed scheme, Mutual Information (MI) extrema are evaluated through an energy constrained objective function, which is based on a-priori information of interference source and desired array factor. Signal to Interference plus Noise Ratio (SINR) performance is evaluated for both transmission and reception. In our scheme, MI is presented as an index to identify trade-off between information gain, SINR, illumination time and spatial selectivity in an energy constrained optimization problem. The employed method yields lesser computational complexity, which is presented through comparative analysis with conventional methods in vogue. MI based beamforming offers enhancement of signal integrity in degraded environment while reducing computational intricacy and correlating key performance indicators.
Lightweight and Seamless Distributed Scheme for the Smart Home
Security of the smart home in terms of behavior activity pattern recognition is a totally dissimilar and unique issue as compared to the security issues of other scenarios. Sensor devices (low capacity and high capacity) interact and negotiate each other by detecting the daily behavior activity of individuals to execute common tasks. Once a device (e.g., surveillance camera, smart phone and light detection sensor etc.) is compromised, an adversary can then get access to a specific device and can damage daily behavior activity by altering the data and commands. In this scenario, a group of common instruction processes may get involved to generate deadlock. Therefore, an effective suitable security solution is required for smart home architecture. This paper proposes seamless distributed Scheme which fortifies low computational wireless devices for secure communication. Proposed scheme is based on lightweight key-session process to upheld cryptic-link for trajectory by recognizing of individual’s behavior activities pattern. Every device and service provider unit (low capacity sensors (LCS) and high capacity sensors (HCS)) uses an authentication token and originates a secure trajectory connection in network. Analysis of experiments is revealed that proposed scheme strengthens the devices against device seizure attack by recognizing daily behavior activities, minimum utilization memory space of LCS and avoids network from deadlock. Additionally, the results of a comparison with other schemes indicate that scheme manages efficiency in term of computation and communication.
Security Issues in Long Term Evolution-Based Vehicle-To-Everything Communication Networks
The ability for vehicles to communicate with other vehicles (V2V), the physical (V2I) and network (V2N) infrastructures, pedestrians (V2P), etc. – collectively known as V2X (Vehicle to Everything) – will enable a broad and growing set of applications and services within the intelligent transport domain for improving road safety, alleviate traffic congestion and support autonomous driving. The telecommunication research and industry communities and standardization bodies (notably 3GPP) has finally approved in Release 14, cellular communications connectivity to support V2X communication (known as LTE – V2X). LTE – V2X system will combine simultaneous connectivity across existing LTE network infrastructures via LTE-Uu interface and direct device-to-device (D2D) communications. In order for V2X services to function effectively, a robust security mechanism is needed to ensure legal and safe interaction among authenticated V2X entities in the LTE-based V2X architecture. The characteristics of vehicular networks, and the nature of most V2X applications, which involve human safety makes it significant to protect V2X messages from attacks that can result in catastrophically wrong decisions/actions include ones affecting road safety. Attack vectors include impersonation attacks, modification, masquerading, replay, MiM attacks, and Sybil attacks. In this paper, we focus our attention on LTE-based V2X security and access control mechanisms. The current LTE-A security framework provides its own access authentication scheme, the AKA protocol for mutual authentication and other essential cryptographic operations between UEs and the network. V2N systems can leverage this protocol to achieve mutual authentication between vehicles and the mobile core network. However, this protocol experiences technical challenges, such as high signaling overhead, lack of synchronization, handover delay and potential control plane signaling overloads, as well as privacy preservation issues, which cannot satisfy the adequate security requirements for majority of LTE-based V2X services. This paper examines these challenges and points to possible ways by which they can be addressed. One possible solution, is the implementation of the distributed peer-to-peer LTE security mechanism based on the Bitcoin/Namecoin framework, to allow for security operations with minimal overhead cost, which is desirable for V2X services. The proposed architecture can ensure fast, secure and robust V2X services under LTE network while meeting V2X security requirements.
Formal Verification for Ethereum Smart Contract Using Coq
The smart contract in Ethereum is a unique program deployed on the Ethereum Virtual Machine (EVM) to help manage cryptocurrency. The security of this smart contract is critical to Ethereum’s operation and highly sensitive. In this paper, we present a formal model for smart contract, using the separated term-obligation (STO) strategy to formalize and verify the smart contract. We use the IBM smart sponsor contract (SSC) as an example to elaborate the detail of the formalizing process. We also propose a formal smart sponsor contract model (FSSCM) and verify SSC’s security properties with an interactive theorem prover Coq. We found the 'Unchecked-Send' vulnerability in the SSC, using our formal model and verification method. Finally, we demonstrate how we can formalize and verify other smart contracts with this approach, and our work indicates that this formal verification can effectively verify the correctness and security of smart contracts.
IT and Security Experts' Innovation and Investment Front for IT-Entrepreneurship in Pakistan
This paper targets the rising factor of entrepreneurship innovation, which lacks in Pakistan as compared to the other countries or the regions like China, India, and Malaysia, etc. This is an exploratory and explanatory study. Major aspects have identified as the direction for the policymakers while highlighting the issues in true spirit. IT needs to be considered not only as a technology but also as itself growing as a new community. IT management processes are complex and broad, so generally requires extensive attention to the collective aspects of human variables, capital and technology. In addition, projects tend to have a special set of critical success factors, and if these are processed and given attention, it will improve the chances of successful implementation. This is only possible with state of the art intelligent decision support systems and accumulating IT staff to some extent in decision processes. This paper explores this issue carefully and discusses six issues to observe the implemented strength and possible enhancement.
Optimized Approach for Secure Data Sharing in Distributed Database
In the current age of technology, information is the most precious asset of a company. Today, companies have a large amount of data. As the data become larger, access to data for some particular information is becoming slower day by day. Faster data processing to shape it in the form of information is the biggest issue. The major problems in distributed databases are the efficiency of data distribution and response time of data distribution. The security of data distribution is also a big issue. For these problems, we proposed a strategy that can maximize the efficiency of data distribution and also increase its response time. This technique gives better results for secure data distribution from multiple heterogeneous sources. The newly proposed technique facilitates the companies for secure data sharing efficiently and quickly.
Requirement Engineering and Software Product Line Scoping Paradigm
Requirement Engineering (RE) is a part being created for programming structure during the software development lifecycle. Software product line development is a new topic area within the domain of software engineering. It also plays important role in decision making and it is ultimately helpful in rising business environment for productive programming headway. Decisions are central to engineering processes and they hold them together. It is argued that better decisions will lead to better engineering. To achieve better decisions requires that they are understood in detail. In order to address the issues, companies are moving towards Software Product Line Engineering (SPLE) which helps in providing large varieties of products with minimum development effort and cost. This paper proposed a new framework for software product line and compared with other models. The results can help to understand the needs in SPL testing, by identifying points that still require additional investigation. In our future scenario, we will combine this model in a controlled environment with industrial SPL projects which will be the new horizon for SPL process management testing strategies.
Web Proxy Detection via Bipartite Graphs and One-Mode Projections
With the Internet becoming the dominant channel for business and life, many IPs are increasingly masked using web proxies for illegal purposes such as propagating malware, impersonate phishing pages to steal sensitive data or redirect victims to other malicious targets. Moreover, as Internet traffic continues to grow in size and complexity, it has become an increasingly challenging task to detect the proxy service due to their dynamic update and high anonymity. In this paper, we present an approach based on behavioral graph analysis to study the behavior similarity of web proxy users. Specifically, we use bipartite graphs to model host communications from network traffic and build one-mode projections of bipartite graphs for discovering social-behavior similarity of web proxy users. Based on the similarity matrices of end-users from the derived one-mode projection graphs, we apply a simple yet effective spectral clustering algorithm to discover the inherent web proxy users behavior clusters. The web proxy URL may vary from time to time. Still, the inherent interest would not. So, based on the intuition, by dint of our private tools implemented by WebDriver, we examine whether the top URLs visited by the web proxy users are web proxies. Our experiment results based on real datasets show that the behavior clusters not only reduce the number of URLs analysis but also provide an effective way to detect the web proxies, especially for the unknown web proxies.
Multi-Dimension Threat Situation Assessment Based on Network Security Attributes
As the increasing network attacks become more and more complex, network situation assessment based on log analysis cannot meet the requirements to ensure network security because of the low quality of logs and alerts. This paper addresses the lack of consideration of security attributes of hosts and attacks in the network. Identity and effectiveness of Distributed Denial of Service (DDoS) are hard to be proved in risk assessment based on alerts and flow matching. This paper proposes a multi-dimension threat situation assessment method based on network security attributes. First, the paper offers an improved Common Vulnerability Scoring System (CVSS) calculation, which includes confident risk, integrity risk, availability risk and a weighted risk. Second, the paper introduces deterioration rate of properties collected by sensors in hosts and network, which aimed at assessing the time and level of DDoS attacks. Third, the paper introduces distribution of asset value in security attributes considering features of attacks and network, which aimed at assessing and show the whole situation. Experiments demonstrate that the approach reflects effectiveness and level of DDoS attacks, and the result can show the primary threat in network and security requirement of network. Through comparison and analysis, the method reflects more in security requirement and security risk situation than traditional methods based on alert and flow analyzing.
Hierarchical Filtering Method of Threat Alerts Based on Correlation Analysis
Nowadays, the threats of the internet are enormous and increasing; however, the classification of huge alert messages generated in this environment is relatively monotonous. It affects the accuracy of the network situation assessment, and also brings inconvenience to the security managers to deal with the emergency. In order to deal with potential network threats effectively and provide more effective data to improve the network situation awareness. It is essential to build a hierarchical filtering method to prevent the threats. In this paper, it establishes a model for data monitoring, which can filter systematically from the original data to get the grade of threats and be stored for using again. Firstly, it filters the vulnerable resources, open ports of host devices and services. Then use the entropy theory to calculate the performance changes of the host devices at the time of the threat occurring and filter again. At last, sort the changes of the performance value at the time of threat occurring. Use the alerts and performance data collected in the real network environment to evaluate and analyze. The comparative experimental analysis shows that the threat filtering method can effectively filter the threat alerts effectively.
An Earth Mover’s Distance Algorithm Based DDoS Detection Mechanism in SDN
Software-defined networking (SDN) provides a solution
for scalable network framework with decoupled control and data
plane. However, this architecture also induces a particular distributed
denial-of-service (DDoS) attack that can affect or even overwhelm
the SDN network. DDoS attack detection problem has to date been
mostly researched as entropy comparison problem. However, this
problem lacks the utilization of SDN, and the results are not accurate.
In this paper, we propose a DDoS attack detection method, which
interprets DDoS detection as a signature matching problem and is
formulated as Earth Mover’s Distance (EMD) model. Considering
the feasibility and accuracy, we further propose to define the cost
function of EMD to be a generalized Kullback-Leibler divergence.
Simulation results show that our proposed method can detect DDoS
attacks by comparing EMD values with the ones computed in the case
without attacks. Moreover, our method can significantly increase the
true positive rate of detection.
An Analysis of the Representation of the Translator and Translation Process into Brazilian Social Networking Groups
In the digital era, in which we have an avalanche of information, it is not new that the Internet has brought new modes of communication and knowledge access. Characterized by the multiplicity of discourses, opinions, beliefs and cultures, the web is a space of political-ideological dimensions where people (who often do not know each other) interact and create representations, deconstruct stereotypes, and redefine identities. Currently, the translator needs to be able to deal with digital spaces ranging from specific software to social media, which inevitably impact on his professional life. One of the most impactful ways of being seen in cyberspace is the participation in social networking groups. In addition to its ability to disseminate information among participants, social networking groups allow a significant personal and social exposure. Such exposure is due to the visibility of each participant achieved not only on its personal profile page, but also in each comment or post the person makes in the groups. The objective of this paper is to study the representations of translators and translation process on the Internet, more specifically in publications in two Brazilian groups of great influence on the Facebook: "Translators/Interpreters" and "Translators, Interpreters and Curious". These chosen groups represent the changes the network has brought to the profession, including the way translators are seen and see themselves. The analyzed posts allowed a reading of what common sense seems to think about the translator as opposed to what the translators seem to think about themselves as a professional class. The results of the analysis lead to the conclusion that these two positions are antagonistic and sometimes represent conflict of interests: on the one hand, the society in general consider the translator’s work something easy, therefore it is not necessary to be well remunerated; on the other hand, the translators who know how complex a translation process is and how much it takes to be a good professional. The results also reveal that social networking sites such as Facebook provide more visibility, but it takes a more active role from the translator to achieve a greater appreciation of the profession and more recognition of the role of the translator, especially in face of increasingly development of automatic translation programs.
A Model for Analysing Argumentative Structures and Online Deliberation in User-Generated Comments to the Website of a South African Newspaper
The conversational dynamics of democratically orientated deliberation continue to stimulate critical scholarship for its potential to bolster robust engagement between different sections of pluralist societies. Several axes of deliberation that have attracted academic attention include face-to-face vs. online interaction, and citizen-to-citizen communication vs. engagement between citizens and political elites. In all these areas, numerous researchers have explored deliberative procedures aimed at achieving instrumental goals such a securing consensus on policy issues, against procedures that prioritise expressive outcomes such as broadening the range of argumentative repertoires that discursively construct and mediate specific political issues. The study that informs this paper, works in the latter stream. Drawing its data from the reader-comments section of a South African broadsheet newspaper, the study investigates online, citizen-to-citizen deliberation by analysing the discursive practices through which competing understandings of social problems are articulated and contested. To advance this agenda, the paper deals specifically with user-generated comments posted in response to news stories on questions of race and racism in South Africa. The analysis works to discern and interpret the various sets of discourse practices that shape how citizens deliberate contentious political issues, especially racism. Since the website in question is designed to encourage the critical comparison of divergent interpretations of news events, without feeding directly into national policymaking, the study adopts an analytic framework that traces how citizens articulate arguments, rather than the instrumental effects that citizen deliberations might exert on policy. The paper starts from the argument that such expressive interactions are particularly crucial to current trends in South African politics, given that the precise nature of race and racism remain contested and uncertain. Centred on a sample of 2358 conversational moves in 814 posts to 18 news stories emanating from issues of race and racism, the analysis proceeds in a two-step fashion. The first stage conducts a qualitative content analysis that offers insights into the levels of reciprocity among commenters (do readers engage with each other or simply post isolated opinions?), as well as the structures of argumentation (do readers support opinions by citing evidence?). The second stage involves a more fine-grained discourse analysis, based on a theorisation of argumentation that delineates it into three components: opinions/conclusions, evidence/data to support opinions/conclusions and warrants that explicate precisely how evidence/data buttress opinions/conclusions. By tracing the manifestation and frequency of specific argumentative practices, this study contributes to the archive of research currently aggregating around the practices that characterise South Africans’ engagement with provocative political questions, especially racism and racial inequity. Additionally, the study also contributes to recent scholarship on the affordances of Web 2.0 software by eschewing a simplistic bifurcation between cyber-optimist vs. pessimism, in favour of a more nuanced and context-specific analysis of the patterns that structure online deliberation.
Quantifying Mobility of Urban Inhabitant Based on Social Media Data
Check-in locations on social media provide information about an individual’s location. The millions of units of data generated from these sites provide knowledge for human activity. In this research, we used a geolocation service and users’ texts posted on Twitter social media to analyze human mobility. Our research will answer the questions; what are the movement patterns of a citizen? And, how far do people travel in the city? We explore the people trajectory of 201,118 check-ins and 22,318 users over a period of one month in Makassar city, Indonesia. To accommodate individual mobility, the authors only analyze the users with check-in activity greater than 30 times. We used sampling method with a systematic sampling approach to assign the research sample. The study found that the individual movement shows a high degree of regularity and intensity in certain places. The other finding found that the average distance an urban inhabitant can travel per day is as far as 9.6 km.
Deploying a Platform as a Service Cloud Solution to Support Student Learning
This presentation describes the design and implementation of PaaS (platform as a service) cloud-based labs that are used in database-related courses to teach students practical skills. Traditionally, all labs are implemented in a desktop-based environment where students have to install heavy client software to access database servers. In order to release students from that burden, we have successfully deployed the cloud-based solution to support database-related courses, from which students and teachers can practice and learn database topics in various database courses via cloud access. With its development environment, execution runtime, web server, database server, and collaboration capability, it offers a shared pool of configurable computing resources and comprehensive environment that supports students’ needs without the complexity of maintaining the infrastructure.
Lexical Based Method for Opinion Detection on Tripadvisor Collection
The massive development of online social networks allows users to post and share their opinions on various topics. With this huge volume of opinion, it is interesting to extract and interpret this information for different domains, e.g., product and service benchmarking, politic, system of recommendation. This is why opinion detection is one of the most important research tasks. It consists of differentiating between opinion data and factual data. The difficulty of this task is to determine an approach which returns opinionated document. Generally, there are two approaches used for opinion detection, i.e., lexical based approaches and machine learning based approaches. In lexical based approaches, a dictionary of sentimental words is used, words are associated with weights. The opinion score of document is derived by the occurrence of words from this dictionary. In machine learning approaches, usually, a classifier is trained using a set of annotated document containing sentiment, and features such as n-grams of words, part-of-speech tags, and logical forms. Majority of these works are based on documents text to determine opinion score but do not take into account if these texts are really correct. Thus, it is interesting to exploit other information to improve opinion detection. In our work, we will develop a new way to consider the opinion score. We introduce the notion of trust score. We determine opinionated documents but also if these opinions are really trustable information in relation with topics. For that we use lexical SentiWordNet to calculate opinion and trust scores, we compute different features about users like (numbers of their comments, numbers of their useful comments, average useful review). After that, we combine opinion score and trust score to obtain a final score. We applied our method to detect trust opinions in TRIPADVISOR collection. Our experimental results report that the combination of opinion score and trust score improves opinion detection.
Study on Construction of 3D Topography by UAV-Based Images
In this paper, a method of fast 3D topography modeling using the high-resolution camera images is studied based on the characteristics of Unmanned Aerial Vehicle (UAV) system for low altitude aerial photogrammetry and the need of three dimensional (3D) urban landscape modeling. Firstly, the existing high-resolution digital camera with special design of overlap images is designed by reconstructing and analyzing the auto-flying paths of UAVs, which improves the self-calibration function to achieve the high precision imaging by software, and further increased the resolution of the imaging system. Secondly, several-angle images including vertical images and oblique images gotten by the UAV system are used for the detail measure of urban land surfaces and the texture extraction. Finally, the aerial photography and 3D topography construction are both developed in campus of Chang-Jung University and in Guerin district area in Tainan, Taiwan, provide authentication model for construction of 3D topography based on combined UAV-based camera images from system. The results demonstrated that the UAV system for low altitude aerial photogrammetry can be used in the construction of 3D topography production, and the technology solution in this paper offers a new, fast, and technical plan for the 3D expression of the city landscape, fine modeling and visualization.
Valuation of Cultural Heritage: A Hedonic Pricing Analysis of Housing via GIS-based Data
The hedonic pricing model has been popularly applied to describe the economic value of environmental amenities in urban housing, but the results for cultural heritage variables remain relatively ambiguous. In this paper, integrated variables extending by GIS-based data and an existing typology of communities used to examine how cultural heritage and environmental amenities and disamenities affect housing prices across urban communities in Tainan, Taiwan. The developed models suggest that, although a sophisticated variable for central services is selected, the centrality of location is not fully controlled in the price models and thus picked up by correlated peripheral and central amenities such as cultural heritage, open space or parks. Analysis of these correlations permits us to qualify results and present a revised set of relatively reliable estimates. Positive effects on housing prices are identified for views, various types of recreational infrastructure and vicinity of nationally cultural sites and significant landscapes. Negative effects are found for several disamenities including wasteyards, refuse incinerators, petrol stations and industries. The results suggest that systematic hypothesis testing and reporting of correlations may contribute to consistent explanatory patterns in hedonic pricing estimates for cultural heritage and landscape amenities in urban.
Information Technology Pattern for Traceability to Increase the Exporting Efficiency of Thailand’s Orchid
Traceability system is one of the tools which can ensure the product’s confident of the consumer as it can trace the product back to its origin and can reduce the operation cost of recall. Nowadays, there are so many technologies which can be applied to the traceability system and also able to increase the efficiency of the system such as QR Code, barcode, GS1 and GTIN. As the result, this research is aimed to study and design the information technology pattern that suits for the traceability of Thailand’s orchid because Thailand’s orchid is the popular export product for Japan, USA, China, Netherlands and Italy. This study will enhance the value of Thailand’s orchid and able to prevent the unexpected event of the defects or damaged product. The traceability pattern was received IOC test from 12 experts from 4 fields of study which are traceability field, information technology field, information communication technology field and orchid export field. The result of the in-depth interview and questionnaire showed that the technology which most compatibility with the traceability system is the QR code. The mean of the score was 4.25 and the standard deviation was 0.5 as the QR code is the new technology and user-friendly. The traceability system should start from the farm to the consumer in the consuming country as the traceability system will enhance the quality level of the product and increase the value of its as well. The other outcome from this research is the supply chain model of Thailand’s Orchid along with the system architecture and working system diagram.