sim.nilim.go.jpsim.nilim.go.jp/Okushiri/download/CPIJ/130823.doc · Web viewAt first, several...

47
3D Archiving Houses and Settlements Alternative Technologies to support Lasting Diachronic Memory KOBAYASHI, Hideyuki (DR. Eng.)* Abstract In order to preserve 3D digital data describing houses, we propose alternative technologies to keep the readability of the data against ever changing data formats, and lasting recording media, through usage of ‘virtual converter’, and ‘electronic record’ attached to each building, as follows: Background and purpose 1. Cases of organizations in charge of maintaining lasting data 2. Historical review of media recording digital data

Transcript of sim.nilim.go.jpsim.nilim.go.jp/Okushiri/download/CPIJ/130823.doc · Web viewAt first, several...

3D Archiving Houses and Settlements

Alternative Technologies to support Lasting Diachronic Memory

KOBAYASHI, Hideyuki (DR. Eng.)*

AbstractIn order to preserve 3D digital data describing houses, we propose alternative technologies to keep the readability of the data against ever changing data formats, and lasting recording media, through usage of ‘virtual converter’, and ‘electronic record’ attached to each building, as follows: Background and purpose

1. Cases of organizations in charge of maintaining lasting data 2. Historical review of media recording digital data 3. Currently used data formats for modeling 3D objects 4. Development of alternative technologies 5. Next steps Conclusion

Keywords: diachronic archiving, 3D data format, recording media, houses and settlements

*Research Coordinator for Housing Information System,Research Center for Advanced Information Technology,National Institute for Land and Infrastructure Management (NILIM), Ministry of Land, Infrastructure, Transport and Tourism (MLIT), JAPANE-Mail: kobayashi-h92qa @ nilim.go .jp

Background and purpose In Japan, housing stock already far exceeds the total number of household, therefore, maintenance of existing houses and buildings became more important, rather than massive construction and provision.

According to the national census on housing and land 2008, 49.99 million households were counted, while total housing stock was 57.59 million units. 7.56million units were identified as ‘empty house’ (fig 1).

Also, the average length of life of houses is extending. There are some different methods for measuring ‘lifetime’, but most precise analysis was performed in Osaka, which showed the ever increasing trend of it.

Figure-2 Trend of average lifetime of timber houses in 2 wards and 1 city in Osaka2)

Average servicetime (Years)

Anno Domino

Higashi-Yodogawa Ward

Hirakata City

, in Osaka Prefecture

Total house unit (left index)

Vacant house units (left index)

Vacant rate (right index)

1958 1963 1968 1973 1978 1983 1988 1993 1998 2003 2008

(million units)

Chuoh WardForecastStatistic

Figure-1 Trend of total house units, vacant units and vacant rate in Japan 1958-20081)

That means, existing houses are expected to serve for generations of families through re-use, or different families through resale of long-life house as stock. This situation is quite different from that after 1945 until 1980s, where new construction and/or ‘scrap and build’ type reconstruction of short-life houses were dominant, which served only for one generation of ‘core families’.

In order to cope with this new situation, the governmental housing policy pays more attention not only to the quality (length of life) of houses, but also provision of reliable data/records of houses attributed to physical buildings will assist the alteration and transactions of in future.

Record of houses and settlements are also valuable in case of disasters. From the technical viewpoint, instead of conventional blue prints and written documents on paper, digital data are becoming popular for synchronic communication. Usefulness of them will further better in future. However, the coded data need media to record and replay, and durability of current digital recording media is still insufficient to cover the whole lifetime of houses. Also, data formats tend to change frequently, making the diachronic transfer of data between ages suspicious. 2D images like drawings and photos could be at least printed on long-life papers and stored. However, coded 3D data cannot be printed on papers as objects of description are seen. They need drives and decoders to be replayed through computer graphics or three dimensional printers etc. This nature of 3D data is common to such coded data as sounds and movies. Therefore, we need durable recording devices, and method for assure the ‘readability’ of various and changing formats of data for the length of time which is, at least, beyond the length of life of physical buildings. In order to strengthen technical basis for to cope with this new demand, ‘Basic Research on Technologies for Permanent Archiving of 3D Data of Houses’ was performed by National Institute for Land and Infrastructure Management (NILIM, in Tsukuba-science city) between fiscal years 2012-2014. This paper reports some alternative technologies we developed, toward the lasting use of 3D data archiving houses and settlements.

1. Cases of organizations in charge of maintaining lasting data Several representative organizations in Japan that are in charge of keeping important and massive digital data were surveyed and technologies, methods and media which they use today for storing ever increasing data were monitored. Many of them are keeping data in own data centers, and carefully considering the out-sourcing of the activities, paying attention to the security of the data.(1) Hokkaido Building Guidance Center The center provides charged service for filing and preserving data of individual houses since 2005, earliest among local authorities in Japan. The houses registered should meet the standard for cold region, which is stricter than the minimum conditions required by the national building standard law. The data can contain digital files (CAD data, image data of drawings, photos etc.) up to 100MB. The data is saved in the data center (out-sourcing to the private firm), and also 3 pieces of CD-R are provided. These three disks are separately filed at the center, at the contractor who built the house and at the owner of the house. The center selects a qualified product of CD-R that is expected to last for more than ten years. When the ten years have passed, the center

sends mails to the contractors and house owners to refresh the media. They say that this renewal procedure is also a good chance to make them recall the existence of such archives. The data center is located in the industrial complex near to Sapporo city, where average temperature is rather low, which is good condition for data center. The data center is also considering the usage of waste heat from the hard discs for regional air conditioning.(2) National Diet Library This organization keeps more than 38 million documents. Most of them are made of paper, however, 670 thousands of recorded media and 8.9 million of documents in the form of micro films are also filed. Besides preservation of original documents, they are elaborating to create digital archives for the purpose of disclosure, without physical damage to the original documents. Sometimes books disclosed are attached with digital data in the form of diskette or CD-ROM. For the purpose of reading, the library keeps machines and drives to read these legacy media.(3) National Archives of Japan Historical documents of the Japanese government are filed in this organization. Important documents are scanned and disclosed in the form of digital archive, accessible through internet. Annex storage is located in Tsukuba, c.a. 50 km from the main storage in Tokyo.(4) Statistical Bureau, Ministry of Internal Affairs National censuses are performed by this organization, and digital master data are coded and filed in their own data center. (5) Geography Survey of Japan

This institute is in charge of geographical survey, especially underground soil conditions and mineral resources, and today under AIST, located in Tsukuba-city.

(6) Geospatial Information Authority of Japan This institute is in charge of geographical survey and publishing various maps. Digital maps (DM) are also available. In 2011, the damaged area of the great eastern Japan earthquake was intensively surveyed and precise maps of changed land shape, damaged settlements were issued.(7) Remote Sensing Technology Center (RESTEC) Satellite image data are collected, kept and delivered from this institute.(8) Building statistic Building certification procedure is performed in Japan under control of Ministry of Land, Infrastructure, Transport and Tourism(MLIT), and newly started construction works are tabulated and issued from the Ministry of Land, Infrastructure, Transport and Tourism. The master data is kept in a governmental data center and also in the form of MO (see 2(4)).(9) Registered copyright of software/programs Copyrights are registered in the Agency for Cultural Affairs, Especially computer programs are registered in Software Information Center (SOFTIC). They had filed the registered works in the form of micro films for lasting memory, however, started to receive digital data since June 2011. Up to march 2013, 11,557 programs have been registered.(10) Public buildings We monitored data on public buildings provided and kept by the local governments. In case of Asahikawa city, CAD data of public offices, schools and rental houses are kept in a file server maintained in the city hall.(11) Technical office of MLIT (SXF) Design drawings and completion drawing of national public works are recorded in a form of

standardized 2D data format named SXF. CAD data of infrastructure provided by consulting firms are converted in the common SXF format and kept in the local technical offices of the MLIT in the form of file server and CD-R, and the data is stored in the local technical offices annex to the 8 provincial branch of the MLIT. They are summarized into GIS and utilized for operation and maintenance works. This GIS data is managed by the NILIM.(12) Japan Patent Office Industrial Property Digital Library (IPDL) keeps digital data of patent and provides searching services. 2. Historical review of media recording digital data Vulnerability and risks of historical digital records are generally warned since 1990s.3)

From this viewpoint, we traced various memory devices which have been used to record digital data (statistics, CAD, GIS etc.) in Building Research Institute (BRI since 1946) and National Institute for Land and Infrastructure Institute (NILIM, established in 2001, partially inherited those old recording media from BRI). They including MT, CMT, FD, MO, DAT, HDD, CD-ROM, CF, SD etc. and we evaluated from viewpoint of durability. In general, devices that keep the data in the form of electrons on silicon chips have the shortest life, followed by magnetic devices. The devices that record the data in the form of physical shape (e.g. punched papers and pressed CD-ROM, not CD-R) can last longest. The durability of listed digital devices promised by the makers has not exceeded 100 years. Beside the deterioration of recording media itself, the social life of the ‘drives’ will be also problematic. Some of the media still correctly keep the recorded data; however, it has become difficult to find the ‘drive’ to read out the data from the media, for example ‘5 inch MO disc’.(1) Paper media for recording program/data In 1970s, holes on paper ribbon or card was used to record program/data. Some of them still remain in our institute for a kind of memory (Figure-3, 4). Through observation, the data recorded on them are still readable through eyes or scanners. However, machines to read them had disappeared (card reader, tape reader, etc.).

Figure-3 One piece of punched card

Figure-4 One series of punched cards containig statistical data

Also, some data printed on papers by ‘line printer’ in 1970s still remains (Figure-5). They seem to be still readable through OCR.

Figure-5 Statistics output by line printers, blueprinted

(2) Magnetic media for recording program/data Open-reel type magnetic tape was a media to keep statistic data in 1960-70s. Large tape drive was equipped in main-frame computer rooms. Magnetic disc was also attached to mini-computers in ‘80s. In those days, open-reel type magnetic tape was still used to record sounds, however the specification was quite different and the tape decks for musical tape are not useful to the tapes containing digital data. In 1970s, compact cassette was introduced to record speech and music. Some of initial micro-computers utilized this for recording and replaying digital data through ‘MODEM’ interface, which was also used to exchange data through telephone. Digital compact cassette was used to record digital data for automation in factories. Tape players for compact cassettes are still produced and sold in the market. DAT, which was once popular for digital recording of audio, was also utilized as back-up media for hard disc in 1990s. Today they are already obsolete. Floppy discs that enable random access between tracks and sectors throughout a disc unit were introduced in BRI (one of previous body of NILIM) in ‘80s. At first, the size was 8 inch, and the density was 250KB. The recording capacity per disc was enhanced through usage of both

sides of a disc and doubling the recording density. In 1990, satellite image data was delivered from RESTEC, in the form of 8inch floppy discs in sequential recording format.

(5 inch double sided double density in MS-DOS format)

Figure-6 One of the most popular types of floppy disk

In order to regulate random access to multiple small files kept in a disc, DOS (disc operation system) was introduced. At first, several DOS’s inherent to various systems (word processor, games etc.) were used, and finally MS-DOS and Macintosh formats became main streams. More compact 5 inches, and 3.5 inches types followed. The last 3.5 inch type had been used until 2010. The 3.5 inch type MS-DOS discs were adopted as official digital media for ‘building certification’ procedure in 1990s. In short, floppy disks had once flourished, and had disappeared rapidly, and during their era (c.a. 20 years), their size, shape, density and format were very varied and rapidly changed, making the future handling very difficult.

Table-2 Types of floppy disks, used in BRI 1980-2000Diameter(inches) Sides Density Capacity Sectors/Track

8 1 Single 250KB 268 2 Double 1.2MB 265 1 Single 160KB 85 1 Double 360KB 95 2 Double 640KB 85 2 Double 720KB 95 2 High 1.2MB 85 2 High 1.44MB 9

3.5 2 Double 640KB 83.5 2 Double 720KB 93.5 2 High 1.2MB 83.5 2 High 1.44MB 9

(4) Optical media for recording program/data CD became popular in ‘80s for delivering recoded sounds, replacing the record discs (SP, then LP). This was also used to deliver programs and data in mass scale, named CD-ROM. In ‘90s, recordable CD-R was introduced to record and exchange large data up to c.a. 650MB. This media is still widely used to keep the CAD data for public buildings and infra. The expected length of life is assumed as 10 years. Production of CR-R media was out-reached to developing countries, resulting cheap and uncertain (no-brand) products. MO was also developed in Japan (5 inch, 3.5inch). 3.5 inch type is still used for recording building statistics. MO was developed in Japan and did not become popular outside Japan; however the quality is still stable. This media is mostly used in governmental offices and private firms today in small amount. The recording density progressed from 160, 230, 640 MB etc., however, the latest drive can still read/write legacy media with lower density. The round recording media is protected in plastic square housing, provided with metal shutter. The recorded data is sensed without physical contact, similar to CD-ROM.

Figure-7 MO disk containing 3D data elaborated in 2000 (3.5 inch, 230MB format)

(5) Electronic media for recording program/data In the beginning of IC memory chips, they were installed on the circuit boards of computers. SRAM and DRAM were memories which forget all after power off. ROM (read only memory, that can contain data without power supply) was also used for keep program for startup the system. UVROM was re-writable, but need to be taken out from the circuit board for programming while power is off. EEPROM can keep the memory during the power is off and was the first re-writable memory on board. Recordable ‘compact flash’ memory, which can keep the memory without power supply and inserted into a standardized slot of machines, was introduced in ‘90s. It was widely used to keep the image data obtained by digital camera. Interface slot/adapters for PC became popular, in order to read the photos taken in fields. The device was also used to exchange the digital data

between PC’s. Soon, several alternative forms with similar chips inside were in sale (SD etc.). Since 2000, similar memory devices provided with USB interface appeared in 2000’s. The capacity of those devices has seemingly extended (20GB, etc), to keep digital photo images with high resolution. However, the durability of the data is not secured. Especially, this type is vulnerable against frequent re-writing access.(6) Data center

Density of HDD (hard disk drives) rapidly developed from 20MB in 1990 to more than 1TB (1,000,000MB) today. The loss of data caused by accidental crush was mitigated by RAID technology through which broken drive can be replaced by new drive while data stored on them is continuously kept and used. Such hard disks connected to local network is widely used to store the data in offices today (file server, NAS etc.).

Since 2001, communal remote data center connected through internet provide service to keep data. Physical device keeping data is hard disc drives, which rapidly enhanced the speed and density. Data center is a kind of out-sourcing of maintenance works for file servers. Data center connected to network is also useful to share the common data among remote users.

However, in case of archiving data of houses and buildings, the demand for access is not frequent. The frequency of access might be only once in ten years. Consumption of electricity is useless to keep such static data. According to a survey performed by a Japanese think tank, total floor area of data center is 1.64 million square meters, and the total power consumption was 7.6 billion kWh in 2010. They also estimated that it will reach 25% of total national power consumption in 20254).

3. Currently used data formats for modeling 3D objects According to our survey, c.a. 300 data formats are used to record 3D objects. Documents on the specification of major representative data formats are collected with sample data and images of the sample objects, as listed in the table 2. Some of them (e.g. VRML) were authorized by ISO; however, other data formats still followed and replaced/added. 3D data is useful to describe the shapes of parts and structure of machines, widely produced by CAD systems and used for CAM production lines.

Table-2 Major 3D data formats surveyed in fy.2010Extension Name

3dm Rhino

3dmf Quickdraw 3D

3ds 3D Studio

3dxml 3D XML

ac AC3D

arc I-DEAS

ase ASCII Scene Export

asm Pro/Engineer, Solid Edge, SolidWorks

atr Lightscape Material

bdl OneSpace Designer

blend Blender

blk Lightscape Blocks

br4 Bryce

c4d Cinema 4D

cab TrueSpace

cadds CADDS

catdrawing, catshape CATIA V5

catpart, catproduct CATIA V5

cgr CATIA Drawing

chr 3Ds Max Characters

dae AutoDesk Collada

ddf Data Descriptive File

dte DTED, National Geospatial-Intelligence Agency (NGA)'s Digital Terrain

Elevation Data

dem Digital Elevation Models

df LightScape Parameter

dlv CATIA V4

drf VIZ Reader

dwf AutoDesk Composer Design Web Format

dwg Legacy AutoCAD Drawing

dws AutoCAD Standards

dwt AutoCAD Drawing Template

dxf AutoCAD Drawing Exchange Format

e00 TIGER/Line Arc Interchange format

eim Electric Image

exp CATIA V4

fac Electric Image

fbx AutoDesk Kaydara FBX

fbl CADfix Log File

fig xfig

flt Flight Studio OpenFlight

fmz FormZ Project File

ft#, rt# TIGER, Topologically Integrated Geographic Encoding and Referencing

gmax AutoDesk Game Creator

geo, scn LSS

gml OpenGIS Geography Markup Language

gts GNU Triangulated Surface

hrc SoftImage

htr Motion Analysis HTR file

ifc Industry Foundation Classes

ige, igs, iges Initial 2D/3D Graphics Exchange Specification

ini POV-Ray animation script

iob 3D Object TDDDB Format

ipt, iam AutoDesk Inventor

iv Open Inventor

jt JT

k3d K-3D Native

kmz Google Earth Model

lay LightScape Layers

lp LightScape Presentation

ls LightScape

lw LightWave 3D

lwo LightWave 3D 5.0 Object

lws LightWave 3D Scene

lxo Luxology Modo

m3g JSR-184

ma Maya Scene ASCII

max 3Ds Max

mb Maya Scene binary

map Quake 3

md2 Quake 2 Player Model

md3 Quake 3

mdd Vertex Key Frame Animation

mel Maya Embedded Language Script

mf1 I-DEAS

model CATIA V4

mot LightWave 3D Motion

mp Maya Scene PLE

ms3d MilkShape 3D

mtx OpenFX Model

ndo Nendo

neu Pro/Engineer

Ntf National Transfer Format

obj Wavefront

obp Bryce

off DEC Object file

par, psm, pwd Solid Edge

pdb PDB Reader v3

pd, pd CADDS

pkg I-DEAS, OneSpace Designer

plt HP-GL

ply Stanford PLY

pov POV-Ray Rendering Instructions

pp2 Poser

prc, prd PRC Adobe 3D Reviewer

prw Adobe 3D Reviewer

prj 3Ds Studio Mesh

prt I-DEAS, NX (Unigraphics), Pro/Engineer

pwc Pulse

pz3 Poser

raw Raw Faces

rib Renderman

rif Radiance

rvt, rte, rfa Rent

rwx Renderware

s3d Strata 3D

sab, sat ACIS

scn TrueSpace

sda, sdp, sds, sdw, ses

sdac, sdpc, sdsc, sdwc

OneSpace Designer

session CATIA V4

sfc OpenGIS Simple Features for CORBA format

shp, shx, dfb, prj Shapefile,  Esri's open hybrid vector data format

skp Google SketchUp

slc, sl, slo Renderman

sldasm, sldlfp, sldprt SolidWorks

slp Pro Engineering

stl Stereo Lithography

stp, step Standard for the Exchange for Product Data

svg Scalable Vector Graphics

tab, dat, id, map MapInfo's vector data format

u3d Universal 3D

unv I-DEAS

vrml Virtual Reality Modeling Language

vue AutoDesk 3D Studio Animation

vw LightScape View

w3d Shockwave 3D Scene

wings Wings 3D

wire Alias Wire

wrl, wrz VRML 1.0, VRML 2.0, VRML 97

x3d, x3dv Extensible 3D (VRML 3.0, uses XML, 2004)

x Direct X

x b, x t, xmt, xmt txt Parasolid

xas, xpr Pro/Engineer

Xyz Simple point cloud formart

BRI also started to utilize 3D data of houses and settlements in 1993 for the purpose of landscape designing. Initial core part of the system ‘Landscape Simulator’ was developed during 1993-96 and started to be delivered through internet as ‘free ware’ since 1996. For this

system, in 1995, a native data format (LSS-G(1)) was defined and programs were coded in C language to input and output in the format, which was packaged in ‘IP’ library (a kind of interpreter for LSS-G commands). The latest version 2.09 was released in March 2011, and modeling functions for creating flat site for housing complexes on the hilly land are added in 2012. However, the data format has not been changed since the original version for 17 years. In the first stage before 1996, modeling functions of the system were very limited, therefore, 3D data of houses and settlements were provided by professional draft man operating commercial CAD system, and then imported to the scene through file converters. Several file converters were elaborated to import external 3D data (Dxf, Microstation, Minicad, Vector Works, VRML, SHP etc.), coded in C language. Modeling of existing houses from photos was also often performed (.skv format). Land shape was also imported through file converter from DEM style data provided by Geography Survey Institute in 2 formats (.mem:50m mesh, .lem:5m mesh). Since beginning of 21th century, mushrooming of new formats occurred. Many of them are XML based. Usually, attributes are defined on 3D objects in various styles. Exporting/importing data of houses and settlements between different systems became inevitable for simulating various issues. Examples are urban scale fire simulation and evacuation in case of emergency, etc. Evaluation of settlements from viewpoint of global environmental issues was also required. In developing countries, CAD software without strict license control was affordable for students, and widely used to produce 3D data. In 2004, in the field of civil engineering, standard data format (SXF) for digital submission of designed data (CAD) was established by NILIM. Result of designing social infrastructure (bridges, tunnels, pavement, banks etc.) are converted into this format, and handed to the branch offices of the ministry. The SXF format specifies only 2D vector data, and 3D format is still under active discussion. However 2D digital data is far easier than the conventional drawings printed on papers for modeling designed 3D objects. Very recently 3D printers became cheap and popular. They require specific data formats. Also, displaying 3D objects including houses as augmented reality (AR) or mixed reality (MR) using popular tablet devices is useful. Usually, adaptation to this situation is performed by applications through providing file I/O functions in the form of ‘plug-in’, or through provision of external files converters among major contemporary file formats. However, from viewpoint of diachronic longevity of data, when a format will become obsolete in future, re-use the legacy data will be difficult. File conversion to contemporary file formats should have been done before the data format is forgotten. If once a file format is forgotten, the data describing an important building cannot be opened. Even an application executable is preserved; the environment to execute the program might be already lost. For example, if we have to re-use a file created by legacy CAD system and saved in an 8 inch floppy disc, double sided high density (2HD), formatted by MS-DOS, then we need following items and steps:(1) Disc drive which can read the media(2) PC to which the drive can be connected(3) OS accompanied by device driver to operate the disc drive (Windows 95 etc.)

(4) Historical CAD system that can open the data file In order to perform all of these, we need a kind of zoo rather than museum to retain antique machines and software at live status. Then we can open the file which is recorded in the diskette (if it is not broken), and observe the shape of building described in the file. However difficulties still remains for us to take out the data in some contemporary file format, and re-use it. If the hardware (PC and disc drive) and software (DOS and CAD system) are not available, then following alternative ways should be taken.(1) Read all the physical tracks and sectors of the floppy disc, neglecting the read-error that may occur, and save it in a long one file in binary format.(2) Analyze the organization of file system (DOS), if file allocation table is not broken.(3) Take out the file of data which describe the desired houses.(4) Provide file converter, based on some historical document on the CAD system at that time, or directly analyze the file format Some of these problems will be solved, if we could describe (2)-(4) in a virtual environment, in which functions of disc drive, dos/OS, cad, display etc could be described in scripts. Virtual zoo of antique machines on virtual machines will be cheaper to maintain than physical zoo of machines and programs. However, if all the information on the format of the data is missing, we need analysis with try and errors. That often occurs also, if the data is given in a format which is popular but not disclosed in an available document (kept secret). In such case, we have to look into the inside of the data. Fortunately, in case of 3D data, the result of the analysis will be proved through the successful 3D presentation on the display by using some contemporary software. The achievement of such archaeological research will be recorded in the form of ‘source code’ of file converters.

It suggests that alternative technologies are needed, which enables keeping data for a long term with intension. The expected term will be longer than the lifetime of houses, with less consumption of energy (electricity), and less risk of unexpected deletion/loss/alteration,. The demand for accessing speed is not the issue, in comparison to the contemporary ‘big data’ on ‘crowd storage’.

Also, the risk of data leak or illegal alteration remains if the data are kept in hard disk drives which are always connected to network.

4. Development of alternative technologies(1) Static and sustainable media for recording In the history of architecture and buildings, record of construction attached to a building played important role to identify the absolute year of construction. Usually, name of carpenters are also recorded on the pentagonal timber plate.

Figure-8 Historical ‘MUNAFUDA’ Figure-9 Electronic recording board, we propose

Even though the amount of data saved on such device is very limited (<1kb), the length of the life of documents written on the timber plate with traditional black ink is quite long. This traditional recording media is called ‘Munafuda’ (a timber plate attached under highest beam of a building recording the date of construction and name of client and carpenters), can last longer than documents on papers. The oldest one dates in 1122 kept in Chuson-ji temple (world heritage) is estimated to have been attached to a lost library building for storing literature of Buddhism5). Documents written on timber plate are called ‘Mokkan’ in general, which are often excavated from ruins after 7th century. Conditions required for archiving 3D data describing houses and settlements are significantly different from that for synchronic transaction/exchange of 3D data for business.- Accessing speed is not essential (several days are allowed to access to the old data)- Frequency of access is very low (maybe once in ten years)- Cost/energy for keeping is important (hopefully free from power supply)- Security against illegal alteration/deletion is important In order to meet these requirements, we chose and proposed ‘curving on long-life material’. Recently fine laser cutters are available, enabling fine (middle dense) recording on lasting materials, like stones. The curving can be performed in small firms. The curved data will be scanned, and the current/future OCR technologies can easily pick up the scanned data. The curved data contains (a) 3D data itself, (b) a metafile to describe the format (for virtual converter), and (c) some additional small images and short literal explanation. (2) Metafile handling to assure readability Since 1980s, efforts to standardize digital data describing buildings and built-up environment have been made. Exchange of data between different CAD systems was desired. Fortunately these problems have been seemingly improved. Useful software that can create 3D models appeared, like SHADE, ArchiCAD, Myhome Designer, Studio 3D Max etc. have been developed. Delivery of 3D data through internet was also enabled through introduction of common DXF, VRML etc. All those efforts were directed to the contemporary data exchange between different systems. If a final ultimate permanent common data format and recording media could be defined and agreed globally, then it would be enough for us to convert any data into it, and store it forever. If it is difficult, then the second choice is to continue efforts to find more useful technologies,

and ever converting the data into the contemporary most popular container at that time. This research tried to find another, the third approach. 3D data is saved in the form of a file, which is equivalent to a bit stream (0110110000), and the preservation of any 3D data is subdivided into 2 issues, namely:a. Media to keep the bit streamb. Format of the bit stream A metafile attached to data is useful to describe how the data describe a house or a settlement. We developed a compiler system, named ‘virtual converter’, which compiles the metafile and create an executable which will input the data file, parsing the grammar, and get elements that describe the house or the settlement.

<Recording phase today>[1] When we preserve a file that record a house, we attach a metafile.[2] We record the set of data file and attached metafile on durable material

<Replaying phase in future> (maybe 100 years later)[3] The bit stream containing data file and metafile is obtained from the recording material[4] The metafile is compiled by virtual converter, creating an executable[5] The executable will open the data file and do actions for usage

The metafile in this system looks like a kind of program for replaying the data file. The metafile need not cover all the specifications of a format (e.g. whole features of dxf) suggested by the extension of data (.dxf) and omit the features that are not referred by the data. However, if some additional dialect or misinterpretation of the specification is performed in the data, it must be followed in the description of metafile, if the feature is significant to preserve. The grammar of the metafile we developed and proposed is a kind of programming language (2)

to be processed by a compiler system. The grammar has following features:[1] A metafile is described by ASCII characters.[2] The grammar of metafile is similar to ANSI-C language, but simpler[3] The metafile can refer to library functions to define 3D data[4] The library functions contains LSS-G commands as subset[5] The library also provides commands for scanning file, and accessing to 3D data in memory Based on this grammar, we developed a new system of ‘virtual converter’, a kind of compiler system, which accepts and processes a pair of 3D data and attached metafile. This compiler will read the metafile and creates a run-time program, which can decode the 3D data in future. The sample applications we provided with the ‘virtual converter’ are (a) file-file converter, (b) virtual reality display, (c) tablet AR viewer (android based) and (d) web based database entry (server based), described in the (3). These are not needed directly when the data is saved. However they are useful to check and debug the metafile to make sure the future readability.

The definition of the grammar (version ID) for creating the metafile must be also attached to the preserved data of each house. However, it is enough to keep the grammar book in e.g. national library, and data for each house refers to the ID of the book.

Even if some of the rules for creating metafiles are changed in future, the revised version of definitions will be stored with different ID, and the change will not affect the ‘readability’ of previous data file and metafile.

Theoretically, the developed method is inspired by the concept of compiler-compiler proposed by Futamura (1971) (3).(3) Prospect of usage in future Of course, we cannot clearly predict the needs and usage of preserved 3D data in future; however, it is worth while assuming some typical usages which are already practical today. For this reason, we developed 4 trial utilizing systems that can input pair of data file and metafile and execute something. All these programs/systems contain the same compilers that handle the pair of data file and metafile, while different linked libraries perform different way of utilization, providing different actions triggered by the same library commands (COORD etc.).(3-1) File-file converter The compiler is triggered by the command line console with three string parameters, which compiles the metafile (first parameter), and the executable (running on the virtual machine) input the data file (second parameter), and finally output a file with the name given as the third parameter. The current version (vc-1c.exe) is developed on VS2005 as console application, and executed on cmd.exe on Windows versions. This version processes any file format defined in the metafile, and library functions write a file in LSS-G format. This simple application was used to develop, test and debug the prototype kernel of the compiler.

Figure-10 Command line (to convert a metafile In_vrml.cmm and a data file house1.wrl into house1.geo)

(3-2) Presentation by virtual reality devices This system receives a pair of metafile and data file, and the obtained 3D building is presented as virtual reality (stereoscopic view). The system compiles the metafile, and the resulted executable input the data file and synthesize the result as 3D objects in the memory of the system, which are presented on the stereoscopic display.

Figure-11 Converting a pair of metafile skv_i.cmm and data file Z(C6_4).geo

1. Title: Data entry to virtual reality presenter 2. Selection of metafile 3. Selection of data file

4. Start operation

Figure-12 Presenting by the VR projector

(3-3) AR viewing in the field The way of usage of preserved 3D data is almost same as (3-2). However this system runs on tablet devices and obtained 3D objects was synthesized on the photo image obtained by an image censor equipped on the back of the device, and the synthesized image is displayed on the LCD of the device. The tested device was ‘Android’ based, instead of ‘Windows’ based (3-2). Therefore, for this system, the source code of ‘virtual converter’, which was developed on VS2005 for Windows, was re-compiled on Android OS by using ‘ndk’ compiler system. The compiled ‘virtual converter’ linked with wrapper application written in Java language run immediately on Android based tablet and almost compatible to that on Windows system(4).

Figure-13, 14 AR view of a pair of metafile and data file by tablet

(3-4) Data storage and conversion on web-server with SQL database This trial system runs on the web-server, with large data storage in SQL server, which contains two different dialogs accessible through network.

Figure-15 Entry point of 3D data storage – web site

1. Title: 3D data storage 2. History of past operation (upload/download) 3. Entry to upload page 4. Entry to download page

A. Upload function Users can assign two local files, namely data file and metafile, and request ‘upload’ function. The server receives two files, and compiles the metafile. The created executable processes the data file and the obtained elements (coordinates, vertexes, lines, faces, solids, etc.) will be stored in the SQL server in the form of tables. This process need far more time than system (3-1)-(3-3). Therefore, the process is performed in the backyard, as a process separate from data accepting procedure. The progress of the process is shown to the client through individual web page created for each upload event. The contents visible at client is automatically updated when the status.

1. Title: 3D data storage 2. Subtitle: Upload

3. Selection of local metafile 4. Selection of local data file

5. Start uploading

Figure-16 Uploading a pair of metafile and data file through web page

B. Download function Users can select uploaded 3D data from the list, and assign a metafile that defines the data format to be downloaded. The executable will select and pick up elements from SQL tables, and synthesize the new 3D file defined by the metafile. This process also take more time equivalent to process A. Therefore, the progress is reported to the users until download is completed. The users can quit the contact to the web server and shutdown the client terminal during the process. Accessing to the page again lets download the result after the completion.

Process A and B are taken at the same time, when a user desire to utilize some preserved old data file and metafile through conversion to contemporary format. However, if desired, the SQL tables created by the upload process can be maintained before the future download process, and SQL tables (physically files on HDD) can be utilized for data storage for a short time (<10 years).

1. Title: 3D data storage 2. Subtitle: Download

3. Selection of 3D data in storage 4. Selection of local metafile

5. Start dowloading

Figure-17 Downloading a data file in any format defined by respective metafile

(4) Possibilities of other alternatives We realized a system of handling data file and metafile, in order to keep the readability of data for a long time. We neither protest that the system we proposed is the only one system, nor the best system to achieve the purpose. Various other systems for the same purpose can be developed. In order to preserve a data, we attach a metafile (1) based on the compiler (1). The definition of the compiler is printed on a book and kept in a library in charge of lasting data preservation, and the future user can compile the metafile and then use the data. Assume that another compiler (2) is developed and another metafile (2) will fulfill the same purpose. We can attach both metafile (1) and metafile (2) to the original data, the future user can use both compiler (1) and compiler (2) as alternatives.

If the compiler (2) performs better than the compiler (1), then those who wish to preserve data will choose compiler (2) to provide metafile, and the users of the compile (1) will decrease. Even in such case, the readability of previous data attached with metafile (1) will be assured in

future. It means that the two compilers (1) and (2) are not competitive, but co-operative.We expect that a metafile is provided and attached to the data file when it is saved and

preserved in the form of durable media. However, later generations can also analyze a data file described in an unknown format and describe the result in the form of a metafile. Since then, the data file will become readable as well as a data file which was attached metafile since beginning. It means that this compiler-metafile system can play a role of tools for archeological studies for digital data.

The most important thing is that at least one compiler exists and is available, to provide a metafile to be attached to the data file.

5. Next steps Technical approaches toward practical production of the developed electronic recording board at appropriate cost will be the main issue. During this experimental stage, sample data will be obtained from past houses that are already lost but requested to be recorded forever and ready to be disclosed.

(1) Indexing past/conventional 2D data into 3D archives Modeling 3D data from old 2D analogue records (mostly papers) and utilize 3D model for searching and indexing original historical records are being challenged. Data of individual houses are collected into settlement model, and the source data for modeling (photos, maps and drawings etc.) will be connected to the 3D model of the lost settlements.(2) Preparing for expected usage in future Simulations will be performed on settlements recalled from archives in desired formats. For that purposes, various metafiles will be elaborated based on the grammar, so that downloading from the server with any desired format for simulating will be obtained.

Conclusion Today, hard disk drive (HDD) is the major device to keep the massive digital data. In order to secure the risk of crush, RAID technologies are utilized, and the out sourcing of data center management was started in the beginning of 21c (called ASP, then ‘cloud’). However, huge consumption of electricity (30Giga Watts to operate the world’s data centers, equivalent to 30 atomic plants) for drives to keep the ever increasing digital data seems to be unsustainable. In order to conserve the data which record past houses and settlements, for future generations with small energy consumption, infrequent access and high security, the method which we developed and proposed in this paper will provide alternatives.

Notes:(1) LSS-G format (1995) is defined in the NILIM’s web site6), and also explained in a printed report7). Library functions available for the Metafile include LSS-G commands as a subset.

(2) Summary of the grammar for the metafile is as follows:・Numerical types: 32 bit integer, 32 bit float, 256 bit quaternion Used for definition of valuables, functions.・Functions Called from other functions

The main function can be omitted, when no sub-functions are used.・Program flow control if(condition){・・・}[else{・・・}] do{・・・}while(condition) while(condition){・・・} for(initialize; condition; incremental{・・・} switch (condition){case }・Library (built-in) functions data scanning from data file data acquisition from SQL server or Memory Data (same command) definition of coordinate, color, vertex definition of line, face definition of significant group of lines or faces (solid body, group of parts)

(3) Futamura8) argued concept of compiler-compiler from viewpoint of partial evaluation in 1971. A program π(C,R)processes constant value C and runtime value R, compiler α converts the program π into α(π,C) (R) which process only the runtime value R, and the process becomes simpler. For example, if π is ‘Y=1+C+R’ where C = 2, then α(π,C) is ‘Y=3+R’. If πis an interpreter int, that executes source code S with runtime data R, the program int(S,R) is

partially evaluated by the compiler to become a programα(int, S)(R) that process only R. (projection 1) Similarly, the relation α(int, S)(R) = α(α, int)(S)(R) is derived, where α(α,int) is a compiler that

translates the source code S to make a runtime program which process the data R. (projection 2) Further, α(α, α)(int) is derived which translates the interpreter ‘int’ into compiler where αfunctions as

‘compiler-compiler’ (projection 3) FILE command in the LSS-G(3) format (1995) for defining models for landscape simulation performs similar. For example, a command G=FILE(B.geo);in an LSS-G file A.geo refers to another file ‘B.geo’, and put a fixed model B as a part of the model A. If the FILE command refers to an executable E, with runtime data R G=FILE(E.exe, R);then the executable E.exe is called with runtime data R and the result of the calculation is put as a part of model A. For example, E.exe will create a door which is half open (parametric definition of a model), and R gives the rate of open, then the fixed shape of a door open 30% will be resulted. In the same way, program C is a converter to import a model F.dxf described in a foreign data format, then G=FILE(C.exe, F.dxf);means a model imported (converted) from F.dxf. (projection 1)In those cases, E.exe or C.exe are native to MS-DOS or Windows environment, and cannot be used on other OS like UNIX or Macintosh. However if VC is a compiler which compiles the source code C.src and create a runtime program whose performance is similar to C.exe, G=FILE(VC, C.src, F.dxf); (projection 2)performs equivalent to

G=FILE(C.exe, F.dxf); on MS-DOS system, and if VC for another operating system is provided, C.src is commonly used on it.

If a common compiler-compiler CC is provided for various operating systems, and the source code of the VC.src is provided, the executable VC is commonly used among those operating systems, and the command will be equivalent to: G=FILE(CC, VC.src, C.src, F.dxf);If we use a multi-platform compiler system CC which is compatible among various operation systems, and define VC.src for that compiler as a permanent agreement as a grammar for metafile C.src, the definition of a converter C.src will be useful forever in order to process F.dxf. G=FILE(CC, GRAMMAR, METAFILE, DATAFILE); (projection 3) Once the compiler-compilers CC native to various platforms process that can generate an executable of virtual converter defined by GRAMMAR, then the executable process METAFILE to create the converter’s executable that input the DATAFILE. However, the purpose of this system is the readability and longevity of digital data, instead of the execution speed which was the basic concept of Futamura.

(4) Minor difference of compilers between VS2005 for Windows and NDK for Android was identified in the handling of address calculation of arrays, minor difference between Windows (VS2005) and Android (ndk) was identified. For example, Integer array and variables

int stack[1000];int stp;

are defined as global variables, and following command is executed.stack [stp] = f1(stack[stp]);If the value of stp changes in the function f1(),

executable created by VS2005 save the return value into the address, using the new stp value after the execution of f1 function, while Android version save it to the address pointed by the stp value before the execution of f1 function. This kind of problems (compatibility) can be solved by extracting the program into two lines:

int stack[1000];int stp, return value;return value = f1(stack[stp]);stack[stp] = return value;

Reference:1) Statistics Bureau. Ministry of Internal Affairs and Communication. “Quick Summary Report of Census on Housing and Land 2008” (in Japanese). http://www.stat.go.jp/data/jyutaku/2008/10_1.htm (accessed 2013-7-1)2) Tsutsumi et-al. A Study of Trend for Stock and Demolition of Wooden Dwelling House. Journal of Architecture and Planning No.649, Architectural Institute of Japan, 2010.5, 695~700 (in Japanese)3) Rothernberg J. ‘Ensuring the Longevity of Digital Information’ 1999, RAND, http://www.clir.org/pubs/archives/ensuring.pdf (accessed 2013-7-1)4) MIC Research Institute. ‘Survey on power consumption by data centers toward green IT’ 2010, in Japanese5) IWATE NIPPO CO., LTD. (A local press company in Iwate prefecture) Memory of past glory of

Hiraizumi: the world heritage (in Japanese). http://www.iwate-np.co.jp/sekai/sekaiisan/sekaiisan4.htm (accessed 2013-7-1).6) Ministry of Construction (1998) Description of Data Structure of Landscape Simulation System (in Japanese). http://sim.nilim.go.jp/VRNCS/MANUALS/Man-STR/INDEX.HTM (accessed 2013-7-1)7) Kobayashi, H. ‘The Architecture of Landscape Simulation System Ver.2.09, provided by MLIT’Research Report of National Institute for Land and Infrastructure Management No.42, (in Japanese), 2011.3, p.53-64, http://www.nilim.go.jp/lab/bcg/siryou/rpn/rpn0042.htm (accessed 2013-7-1)8) Futamura, Y. “Partial Evaluation of Computation Process-an Approach to a Compiler-Compiler-“, Systems, Computers, Controls" Vol.2, No.5, 1971, p45-50