P
polisci grad
I am currently working on developing a database, in Access 2003, to be used
by mulitple users (albeit in a small environment) who will not be adding or
changing data. I have little experience in Access, although I have read
through many of the tutorials, training sessions, and i have a big book on
the stuff too. However, I am concerned about the way i have normalized my
tables.
this database is to consist of a large amount of disparate and unrelated
data. it is to serve an organization involved in the transportation
industry, but the range of valuable data is huge. for instance, some of my
variables include km of road, # of road collisions, # of vehicle sales, % of
industry share, and so on.
As such, I have approached the problem by constructing a large main table
for all raw data (named data) and small tables to define coded values (like 1
for safety). My fields are Generated ID, Theme (ranging from 1-7), Region
(from 1-20), Industry (1-6), Year (1990-2006), Unit (abbv. title for
variable), and value_#, value_$, value_%, explanation, source, and notes. as
such, each number gets its own record, meaning a simple table in excel
becomes many lines of similar looking data.
now, since one 'unit' or variable may contain value_# (numeric data) for
various regions, years, industries, the amount of repeated data is large,
begging the question of proper normalization. while i have managed to
construct some basic queries, i am worried i may have limited further
advancement with this construct. can anybody offer suggestions based on my
short description of the problem?
by mulitple users (albeit in a small environment) who will not be adding or
changing data. I have little experience in Access, although I have read
through many of the tutorials, training sessions, and i have a big book on
the stuff too. However, I am concerned about the way i have normalized my
tables.
this database is to consist of a large amount of disparate and unrelated
data. it is to serve an organization involved in the transportation
industry, but the range of valuable data is huge. for instance, some of my
variables include km of road, # of road collisions, # of vehicle sales, % of
industry share, and so on.
As such, I have approached the problem by constructing a large main table
for all raw data (named data) and small tables to define coded values (like 1
for safety). My fields are Generated ID, Theme (ranging from 1-7), Region
(from 1-20), Industry (1-6), Year (1990-2006), Unit (abbv. title for
variable), and value_#, value_$, value_%, explanation, source, and notes. as
such, each number gets its own record, meaning a simple table in excel
becomes many lines of similar looking data.
now, since one 'unit' or variable may contain value_# (numeric data) for
various regions, years, industries, the amount of repeated data is large,
begging the question of proper normalization. while i have managed to
construct some basic queries, i am worried i may have limited further
advancement with this construct. can anybody offer suggestions based on my
short description of the problem?