MAPREDUCE-3171. normalize nodemanager native code compilation with common/hdfs native. (tucu)

git-svn-id: https://svn.apache.org/repos/asf/hadoop/common/trunk@1182575 13f79535-47bb-0310-9956-ffa450edef68
This commit is contained in:
Alejandro Abdelnur 2011-10-12 20:38:51 +00:00
parent f143895bc2
commit 327e02fdf4
14 changed files with 44 additions and 80 deletions

View File

@ -9,11 +9,8 @@
<!-- TODO: this layout is wrong. We need module specific bin files in module specific dirs --> <!-- TODO: this layout is wrong. We need module specific bin files in module specific dirs -->
<fileSets> <fileSets>
<fileSet> <fileSet>
<directory>hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/target/classes/bin</directory> <directory>hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/target/native/target/usr/local/bin</directory>
<outputDirectory>bin</outputDirectory> <outputDirectory>bin</outputDirectory>
<includes>
<include>container-executor</include>
</includes>
<fileMode>0755</fileMode> <fileMode>0755</fileMode>
</fileSet> </fileSet>
<fileSet> <fileSet>
@ -54,7 +51,6 @@
<exclude>**/SecurityAuth.audit*</exclude> <exclude>**/SecurityAuth.audit*</exclude>
</excludes> </excludes>
<includes> <includes>
<include>assembly/**</include>
<include>pom.xml</include> <include>pom.xml</include>
<include>build*.xml</include> <include>build*.xml</include>
<include>ivy.xml</include> <include>ivy.xml</include>

View File

@ -22,6 +22,9 @@ Trunk (unreleased changes)
MAPREDUCE-2836. Provide option to fail jobs when submitted to non-existent MAPREDUCE-2836. Provide option to fail jobs when submitted to non-existent
fair scheduler pools. (Ahmed Radwan via todd) fair scheduler pools. (Ahmed Radwan via todd)
MAPREDUCE-3171. normalize nodemanager native code compilation with common/hdfs
native. (tucu)
BUG FIXES BUG FIXES
MAPREDUCE-3166. [Rumen] Make Rumen use job history api instead of relying MAPREDUCE-3166. [Rumen] Make Rumen use job history api instead of relying

View File

@ -47,72 +47,43 @@
<version>1.0-beta-1</version> <version>1.0-beta-1</version>
<executions> <executions>
<execution> <execution>
<id>clean</id> <id>compile</id>
<phase>clean</phase>
<configuration>
<arguments>
<argument>-i</argument>
</arguments>
<workDir>src/main/c/container-executor</workDir>
<destDir>target</destDir>
</configuration>
<goals>
<goal>autoreconf</goal>
<goal>make-clean</goal>
</goals>
</execution>
<execution>
<id>make</id>
<phase>compile</phase> <phase>compile</phase>
<configuration>
<workDir>src/main/c/container-executor</workDir>
<configureEnvironment>
<property>
<name>CFLAGS</name>
<value>-DHADOOP_CONF_DIR=${container-executor.conf.dir}</value>
</property>
</configureEnvironment>
<sources>
<source>
<directory>src/main/c/container-executor</directory>
</source>
</sources>
<workDir>src/main/c/container-executor</workDir>
<destDir>target</destDir>
<prefix>${project.build.outputDirectory}</prefix>
</configuration>
<goals> <goals>
<goal>autoreconf</goal> <goal>autoreconf</goal>
<goal>configure</goal> <goal>configure</goal>
<goal>compile</goal> <goal>make-install</goal>
</goals> </goals>
</execution> </execution>
<execution> <execution>
<id>make-test</id> <id>test</id>
<phase>test</phase> <phase>test</phase>
<configuration>
<workDir>src/main/c/container-executor</workDir>
<configureEnvironment>
<property>
<name>CFLAGS</name>
<value>-DHADOOP_CONF_DIR=${container-executor.conf.dir}</value>
</property>
</configureEnvironment>
<sources>
<source>
<directory>src/main/c/container-executor</directory>
</source>
</sources>
<workDir>src/main/c/container-executor</workDir>
<destDir>target</destDir>
<prefix>${project.build.outputDirectory}</prefix>
<compileTarget>check</compileTarget>
</configuration>
<goals> <goals>
<goal>compile</goal> <goal>test</goal>
</goals> </goals>
</execution> </execution>
</executions> </executions>
<configuration>
<!-- autoreconf settings -->
<workDir>${project.build.directory}/native/container-executor</workDir>
<arguments>
<argument>-i</argument>
</arguments>
<!-- configure settings -->
<configureEnvironment>
<property>
<name>CFLAGS</name>
<value>-DHADOOP_CONF_DIR=${container-executor.conf.dir}</value>
</property>
</configureEnvironment>
<configureWorkDir>${project.build.directory}/native/container-executor</configureWorkDir>
<prefix>/usr/local</prefix>
<!-- configure & make settings -->
<destDir>${project.build.directory}/native/target</destDir>
</configuration>
</plugin> </plugin>
</plugins> </plugins>
</build> </build>
@ -191,6 +162,21 @@
<goal>run</goal> <goal>run</goal>
</goals> </goals>
</execution> </execution>
<execution>
<id>compile</id>
<phase>generate-sources</phase>
<goals>
<goal>run</goal>
</goals>
<configuration>
<target>
<mkdir dir="${project.build.directory}/native"/>
<copy toDir="${project.build.directory}/native">
<fileset dir="${basedir}/src/main/native"/>
</copy>
</target>
</configuration>
</execution>
</executions> </executions>
</plugin> </plugin>

View File

@ -1,21 +0,0 @@
Makefile
Makefile.in
aclocal.m4
compile
config.log
config.status
configure
depcomp
impl/.deps/
impl/.dirstamp
impl/configuration.o
impl/main.o
impl/container-executor.o
install-sh
libtool
missing
container-executor
test/.deps/
test-task-controller
test/.dirstamp
test/test-task-controller.o