This list is closed, nobody may subscribe to it.
2010 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
(139) |
Aug
(94) |
Sep
(232) |
Oct
(143) |
Nov
(138) |
Dec
(55) |
---|---|---|---|---|---|---|---|---|---|---|---|---|
2011 |
Jan
(127) |
Feb
(90) |
Mar
(101) |
Apr
(74) |
May
(148) |
Jun
(241) |
Jul
(169) |
Aug
(121) |
Sep
(157) |
Oct
(199) |
Nov
(281) |
Dec
(75) |
2012 |
Jan
(107) |
Feb
(122) |
Mar
(184) |
Apr
(73) |
May
(14) |
Jun
(49) |
Jul
(26) |
Aug
(103) |
Sep
(133) |
Oct
(61) |
Nov
(51) |
Dec
(55) |
2013 |
Jan
(59) |
Feb
(72) |
Mar
(99) |
Apr
(62) |
May
(92) |
Jun
(19) |
Jul
(31) |
Aug
(138) |
Sep
(47) |
Oct
(83) |
Nov
(95) |
Dec
(111) |
2014 |
Jan
(125) |
Feb
(60) |
Mar
(119) |
Apr
(136) |
May
(270) |
Jun
(83) |
Jul
(88) |
Aug
(30) |
Sep
(47) |
Oct
(27) |
Nov
(23) |
Dec
|
2015 |
Jan
|
Feb
|
Mar
|
Apr
|
May
|
Jun
|
Jul
|
Aug
|
Sep
(3) |
Oct
|
Nov
|
Dec
|
2016 |
Jan
|
Feb
|
Mar
(4) |
Apr
(1) |
May
|
Jun
|
Jul
|
Aug
|
Sep
|
Oct
|
Nov
|
Dec
|
From: <tob...@us...> - 2014-05-06 18:56:09
|
Revision: 8209 http://sourceforge.net/p/bigdata/code/8209 Author: tobycraig Date: 2014-05-06 18:56:06 +0000 (Tue, 06 May 2014) Log Message: ----------- Improved detection of inserted content in update panel Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-06 17:57:47 UTC (rev 8208) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-06 18:56:06 UTC (rev 8209) @@ -310,7 +310,7 @@ // when file:// is present, Windows paths must use / and may include a : // http[s]:// is mandatory for HTTP paths var unix = /^(file:\/\/)?((\/[^\/]+)+)$/; - var windows = /^((file:\/\/)([A-Za-z]:?([\/][^\/\\]+)+))|([A-Za-z]:([\\\/][^\\\/]+)+)$/; + var windows = /^(((file:\/\/)([A-Za-z]:?([\/][^\/\\]+)+))|([A-Za-z]:([\\\/][^\\\/]+)+))$/; var http = /^https?:\/((\/[^\/]+)+)$/; if(unix.test(text.trim()) || windows.test(text.trim()) || http.test(text.trim())) { return 'path'; @@ -374,7 +374,7 @@ 'trix': 'application/trix', 'turtle': 'application/x-turtle'}; -var sparql_update_commands = ['INSERT', 'DELETE']; +var sparql_update_commands = ['INSERT', 'DELETE', 'LOAD', 'CLEAR']; $('#update-file').change(handleFile); $('#update-box').on('dragover', handleDragOver) This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <dme...@us...> - 2014-05-06 17:57:50
|
Revision: 8208 http://sourceforge.net/p/bigdata/code/8208 Author: dmekonnen Date: 2014-05-06 17:57:47 +0000 (Tue, 06 May 2014) Log Message: ----------- Updates to sync with new jetty.xml and other updates through April. EC2 instance reboots can now be avoided (verification TBD). Modified Paths: -------------- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Berksfile branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Vagrantfile branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/attributes/default.rb branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/bin/setHosts.py branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/metadata.rb branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/default/bigdataHA.erb branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/jetty.xml.erb branches/DEPLOYMENT_BRANCH_1_3_1/src/resources/etc/init.d/bigdataHA Removed Paths: ------------- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/init.d/bigdataHA.erb Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Berksfile =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Berksfile 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Berksfile 2014-05-06 17:57:47 UTC (rev 8208) @@ -7,6 +7,6 @@ cookbook 'lvm' cookbook "hadoop" cookbook "emacs" -# cookbook "ganglia" +cookbook "sysstat" metadata Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Vagrantfile =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Vagrantfile 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/Vagrantfile 2014-05-06 17:57:47 UTC (rev 8208) @@ -63,6 +63,7 @@ "recipe[apt]", "recipe[java]", "recipe[emacs]", # default is no x11 + "recipe[sysstat]", # # use this recipe only if the oracle server is timing out in the above 'java' recipe # @@ -135,6 +136,7 @@ "recipe[apt]", "recipe[java]", "recipe[emacs]", # default is no x11 + "recipe[sysstat]", # # use this recipe only if the oracle server is timing out in the above 'java' recipe # @@ -207,6 +209,7 @@ "recipe[apt]", "recipe[java]", "recipe[emacs]", # default is no x11 + "recipe[sysstat]", # # use this recipe only if the oracle server is timing out in the above 'java' recipe # Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/attributes/default.rb =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/attributes/default.rb 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/attributes/default.rb 2014-05-06 17:57:47 UTC (rev 8208) @@ -16,7 +16,7 @@ default['systap-bigdataHA'][:data_dir] = node['systap-bigdataHA'][:fed_dir] + "/data" # Where the log files will live: -default['systap-bigdataHA'][:log_dir] = node['systap-bigdataHA'][:fed_dir] + "/logs" +default['systap-bigdataHA'][:log_dir] = node['systap-bigdataHA'][:fed_dir] + "/log" # Name of the replication cluster to which this HAJournalServer will belong. default['systap-bigdataHA'][:logical_service_id] = 'HA-Replication-Cluster-1' Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/bin/setHosts.py =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/bin/setHosts.py 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/bin/setHosts.py 2014-05-06 17:57:47 UTC (rev 8208) @@ -35,8 +35,12 @@ def createZookeeperSubstitution( index, host, ipAddress ): return "sudo sed -i 's|server." + index + "=" + host + "|server." + index + "=" + ipAddress + "|' /etc/zookeeper/conf/zoo.cfg" -def createBigdataHASubstitution( host, ipAddress ): - return "sudo sed -i 's|" + host + "|" + ipAddress + "|' /etc/defaults/bigdataHA" +def createJiniLocatorsSubstitution(): + locators = "" + for host in hostMap: + locators = locators + "jini://" + hostMap[host] + "/," + locators = locators[:-1] + return "sudo sed -i 's|%JINI_LOCATORS%|" + locators + "|' /etc/default/bigdataHA" if __name__ == '__main__': @@ -55,6 +59,9 @@ private_security_group_name = os.environ["AWS_SECURITY_GROUP_PRIVATE"] group = ec2conn.get_all_security_groups( private_security_group_name )[0] + jini_locators = createJiniLocatorsSubstitution() + # print "JINI_LOCATORS = " + jini_locators + i = 1 for host in bigdataHosts: ssh_client = sshclient_from_instance( host, key_path, user_name='ubuntu' ) @@ -70,11 +77,9 @@ status, stdin, stderr = ssh_client.run( createZookeeperSubstitution( "2", bigdataB, hostMap[ bigdataB ] ) ) status, stdin, stderr = ssh_client.run( createZookeeperSubstitution( "3", bigdataC, hostMap[ bigdataC ] ) ) - name = host.__dict__['tags']['Name'] - hostAddress = host.__dict__['private_ip_address'] - # status, stdin, stderr = ssh_client.run( createBigdataHASubstitution( name, hostAddress ) ) + status, stdin, stderr = ssh_client.run( jini_locators ) - hostAddress = hostAddress + "/32" + hostAddress = host.__dict__['private_ip_address'] + "/32" group.authorize( ip_protocol="tcp", from_port="0", to_port="65535", cidr_ip=hostAddress, src_group=None ) i += 1 @@ -88,4 +93,4 @@ print "The hosts are now rebooting, this may take several minutes. \nOnce back up, you may confirm status by visiting:\n" for host in bigdataHosts: - print "\thttp://" + host.__dict__['ip_address'] + ":8080/status\n" + print "\thttp://" + host.__dict__['ip_address'] + ":8080/bigdata/status\n" Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/metadata.rb =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/metadata.rb 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/metadata.rb 2014-05-06 17:57:47 UTC (rev 8208) @@ -13,3 +13,4 @@ depends 'lvm' depends 'hadoop' depends 'emacs' +depends 'sysstat' Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/default/bigdataHA.erb =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/default/bigdataHA.erb 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/default/bigdataHA.erb 2014-05-06 17:57:47 UTC (rev 8208) @@ -32,7 +32,7 @@ # Apache River - NO default for "LOCATORS". export GROUPS="${FEDNAME}" -export LOCATORS="jini://<%= node['systap-bigdataHA'][:river_locator1] %>/,jini://<%= node['systap-bigdataHA'][:river_locator2] %>/,jini://<%= node['systap-bigdataHA'][:river_locator3] %>/" +export LOCATORS="%JINI_LOCATORS%" # Apache ZooKeeper - NO default. export ZK_SERVERS="<%= node['systap-bigdataHA'][:zk_server1] %>:<%= node['zookeeper'][:zoocfg][:clientPort] %>,<%= node['systap-bigdataHA'][:zk_server2] %>:<%= node['zookeeper'][:zoocfg][:clientPort] %>,<%= node['systap-bigdataHA'][:zk_server3] %>:<%= node['zookeeper'][:zoocfg][:clientPort] %>" Deleted: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/init.d/bigdataHA.erb =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/init.d/bigdataHA.erb 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/init.d/bigdataHA.erb 2014-05-06 17:57:47 UTC (rev 8208) @@ -1,133 +0,0 @@ -#!/bin/bash - -# init.d style script for bigdata HA services. The script can be used -# to 'start' or 'stop' services. -# -# Environment: -# -# binDir - The directory containing the installed scripts. -# pidFile - The pid is written on this file. -# -# Misc. -# -# See http://tldp.org/LDP/abs/html/index.html -# -# Note: Blank lines are significant in shell scripts. -# -# Note: Children must do "exit 0" to indicate success. -# -# Note: Convert DOS cr-lf to unix style in emacs: C-x RET f then unix - -# Source function library (just used for 'action'). If you don't have this -# it SHOULD automatically use the inline definition for "action()". -if [ -f "/etc/init.d/functions" ]; then - . /etc/init.d/functions -else -# Run some action. Log its output. No fancy colors. First argument is the -# label for the log file. Remaining arguments are the command to execute -# and its arguments, if any. - action() { - local STRING rc - STRING=$1 - echo -n "$STRING " - shift - sudo -u bigdata -g bigdata "$@" && echo -n "[OK]" || echo -n "[FAILED]" - rc=$? - echo - return $rc - } -fi - -# Where the scripts live. -cd `dirname $0` - -## -# Highly Recommended OS Tuning. -## - -# Do not swap out applications while there is free memory. -#/sbin/sysctl -w vm.swappiness=0 - -# Setup the environment. -source /etc/default/bigdataHA - -if [ -z "$binDir" ]; then - echo $"$0 : environment not setup: binDir is undefined." - exit 1; -fi -if [ -z "$pidFile" ]; then - echo $"$0 : environment not setup: pidFile is undefined" - exit 1; -fi - -# -# See how we were called. -# -case "$1" in - start) - - cd <%= node['systap-bigdataHA'][:fed_dir] %> -# -# Start the ServiceStarter and child services if not running. -# - if [ -f "$pidFile" ]; then - read pid < "$pidFile" - pidno=$( ps ax | grep $pid | awk '{ print $1 }' | grep $pid ) - if [ -z "$pidno" ]; then -# The process has died so remove the old pid file. - echo $"`date` : `hostname` : $pid died?" - rm -f "$pidFile" - fi - fi - if [ ! -f "$pidFile" ]; then - action $"`date` : `hostname` : bringing up services: " $binDir/startHAServices - else - echo $"`date` : `hostname` : running as $pid" - fi - ;; - stop) -# -# Stop the ServiceStarter and all child services. -# - if [ -f "$pidFile" ]; then - read pid < "$pidFile" - pidno=$( ps ax | grep $pid | awk '{ print $1 }' | grep $pid ) - if [ -z "$pidno" ]; then -# The process has died so remove the old pid file. - echo $"`date` : `hostname` : $pid died?" - rm -f "$pidFile" - else - action $"`date` : `hostname` : bringing down services: " kill $pid - rm -f "$pidFile" - fi - fi - ;; - status) -# -# Report status for the ServicesManager (up or down). -# - if [ -f "$pidFile" ]; then - read pid < "$pidFile" - pidno=$( ps ax | grep $pid | awk '{ print $1 }' | grep $pid ) - if [ -z "$pidno" ]; then - echo $"`date` : `hostname` : process died? pid=$pid." - else - echo $"`date` : `hostname` : running as $pid." - fi - else - echo $"`date` : `hostname` : not running." - fi - ;; - restart) - $0 stop - $0 start - ;; - *) -# -# Usage -# - echo $"Usage: $0 {start|stop|status|restart}" - exit 1 -esac - -exit 0 Modified: branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/jetty.xml.erb =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/jetty.xml.erb 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/bigdata/src/resources/deployment/vagrant/systap-aws-bigdata-ha/templates/default/jetty.xml.erb 2014-05-06 17:57:47 UTC (rev 8208) @@ -5,29 +5,119 @@ <Configure id="Server" class="org.eclipse.jetty.server.Server"> <!-- =========================================================== --> - <!-- Server Thread Pool --> + <!-- Configure the Server Thread Pool. --> + <!-- The server holds a common thread pool which is used by --> + <!-- default as the executor used by all connectors and servlet --> + <!-- dispatches. --> + <!-- --> + <!-- Configuring a fixed thread pool is vital to controlling the --> + <!-- maximal memory footprint of the server and is a key tuning --> + <!-- parameter for tuning. In an application that rarely blocks --> + <!-- then maximal threads may be close to the number of 5*CPUs. --> + <!-- In an application that frequently blocks, then maximal --> + <!-- threads should be set as high as possible given the memory --> + <!-- available. --> + <!-- --> + <!-- Consult the javadoc of o.e.j.util.thread.QueuedThreadPool --> + <!-- for all configuration that may be set here. --> <!-- =========================================================== --> - <Set name="ThreadPool"> - <!-- Default queued blocking threadpool --> - <New class="org.eclipse.jetty.util.thread.QueuedThreadPool"> - <Set name="minThreads">10</Set> - <Set name="maxThreads">64</Set> - </New> - </Set> + <!-- uncomment to change type of threadpool --> + <Arg name="threadpool"><New id="threadpool" class="org.eclipse.jetty.util.thread.QueuedThreadPool"/></Arg> + <!-- --> + <Get name="ThreadPool"> + <Set name="minThreads" type="int"><SystemProperty name="jetty.threads.min" default="10"/></Set> + <Set name="maxThreads" type="int"><SystemProperty name="jetty.threads.max" default="64"/></Set> + <Set name="idleTimeout" type="int"><SystemProperty name="jetty.threads.timeout" default="60000"/></Set> + <Set name="detailedDump">false</Set> + </Get> + <!-- =========================================================== --> + <!-- Get the platform mbean server --> + <!-- =========================================================== --> + <Call id="MBeanServer" class="java.lang.management.ManagementFactory" + name="getPlatformMBeanServer" /> + + <!-- =========================================================== --> + <!-- Initialize the Jetty MBean container --> + <!-- =========================================================== --> + <Call name="addBean"> + <Arg> + <New id="MBeanContainer" class="org.eclipse.jetty.jmx.MBeanContainer"> + <Arg> + <Ref refid="MBeanServer" /> + </Arg> + </New> + </Arg> + </Call> + + <!-- Add the static log to the MBean server. --> + <Call name="addBean"> + <Arg> + <New class="org.eclipse.jetty.util.log.Log" /> + </Arg> + </Call> + + <!-- For remote MBean access (optional) --> + <New id="ConnectorServer" class="org.eclipse.jetty.jmx.ConnectorServer"> + <Arg> + <New class="javax.management.remote.JMXServiceURL"> + <Arg type="java.lang.String">rmi</Arg> + <Arg type="java.lang.String" /> + <Arg type="java.lang.Integer"><SystemProperty name="jetty.jmxrmiport" default="1090"/></Arg> + <Arg type="java.lang.String">/jndi/rmi://<SystemProperty name="jetty.jmxrmihost" default="localhost"/>:<SystemProperty name="jetty.jmxrmiport" default="1099"/>/jmxrmi</Arg> + </New> + </Arg> + <Arg>org.eclipse.jetty.jmx:name=rmiconnectorserver</Arg> + <Call name="start" /> + </New> + <!-- =========================================================== --> - <!-- Set connectors --> + <!-- Http Configuration. --> + <!-- This is a common configuration instance used by all --> + <!-- connectors that can carry HTTP semantics (HTTP, HTTPS, SPDY)--> + <!-- It configures the non wire protocol aspects of the HTTP --> + <!-- semantic. --> + <!-- --> + <!-- Consult the javadoc of o.e.j.server.HttpConfiguration --> + <!-- for all configuration that may be set here. --> <!-- =========================================================== --> + <New id="httpConfig" class="org.eclipse.jetty.server.HttpConfiguration"> + <Set name="secureScheme">https</Set> + <Set name="securePort"><Property name="jetty.secure.port" default="8443" /></Set> + <Set name="outputBufferSize"><Property name="jetty.output.buffer.size" default="32768" /></Set> + <Set name="requestHeaderSize"><Property name="jetty.request.header.size" default="8192" /></Set> + <Set name="responseHeaderSize"><Property name="jetty.response.header.size" default="8192" /></Set> + <Set name="sendServerVersion"><Property name="jetty.send.server.version" default="true" /></Set> + <Set name="sendDateHeader"><Property name="jetty.send.date.header" default="false" /></Set> + <Set name="headerCacheSize">512</Set> + <!-- Uncomment to enable handling of X-Forwarded- style headers + <Call name="addCustomizer"> + <Arg><New class="org.eclipse.jetty.server.ForwardedRequestCustomizer"/></Arg> + </Call> + --> + </New> + + <!-- Configure the HTTP endpoint. --> + <Call name="addConnector"> + <Arg> + <New class="org.eclipse.jetty.server.ServerConnector"> + <Arg name="server"><Ref refid="Server" /></Arg> + <Arg name="factories"> + <Array type="org.eclipse.jetty.server.ConnectionFactory"> + <Item> + <New class="org.eclipse.jetty.server.HttpConnectionFactory"> + <Arg name="config"><Ref refid="httpConfig" /></Arg> + </New> + </Item> + </Array> + </Arg> + <Set name="host"><Property name="jetty.host" /></Set> + <Set name="port"><Property name="jetty.port" default="8080" /></Set> + <Set name="idleTimeout"><Property name="http.timeout" default="30000"/></Set> + </New> + </Arg> + </Call> - <Call name="addConnector"> - <Arg> - <New class="org.eclipse.jetty.server.nio.SelectChannelConnector"> - <Set name="host"><Property name="jetty.host" /></Set> - <Set name="port"><Property name="jetty.port" default="8080"/></Set> - </New> - </Arg> - </Call> - <!-- =========================================================== --> <!-- Set handler Collection Structure --> <!-- =========================================================== --> @@ -36,25 +126,29 @@ <Set name="handlers"> <Array type="org.eclipse.jetty.server.Handler"> <Item> + <!-- This is the bigdata web application. --> <New id="WebAppContext" class="org.eclipse.jetty.webapp.WebAppContext"> - <!-- The location of the top-level of the bigdata webapp. --> - <Set name="resourceBase"> - <Property name="jetty.resourceBase" default="<%= node['systap-bigdataHA'][:fed_dir] %>/var/jetty" /> + <Set name="war"> + <!-- The location of the top-level of the bigdata webapp. --> + <Property name="jetty.resourceBase" default="<%= node['systap-bigdataHA'][:fed_dir] %>/var/jetty" /> </Set> - <Set name="contextPath">/</Set> + <Set name="contextPath">/bigdata</Set> <Set name="descriptor"><%= node['systap-bigdataHA'][:fed_dir] %>/var/jetty/WEB-INF/web.xml</Set> <Set name="parentLoaderPriority">true</Set> <Set name="extractWAR">false</Set> - <Set name="welcomeFiles"> - <Array type="java.lang.String"> - <Item>index.html</Item> - </Array> - </Set> </New> </Item> </Array> </Set> </New> </Set> - + + <!-- =========================================================== --> + <!-- extra server options --> + <!-- =========================================================== --> + <Set name="stopAtShutdown">true</Set> + <Set name="stopTimeout">5000</Set> + <Set name="dumpAfterStart"><Property name="jetty.dump.start" default="false"/></Set> + <Set name="dumpBeforeStop"><Property name="jetty.dump.stop" default="false"/></Set> + </Configure> Modified: branches/DEPLOYMENT_BRANCH_1_3_1/src/resources/etc/init.d/bigdataHA =================================================================== --- branches/DEPLOYMENT_BRANCH_1_3_1/src/resources/etc/init.d/bigdataHA 2014-05-06 13:54:41 UTC (rev 8207) +++ branches/DEPLOYMENT_BRANCH_1_3_1/src/resources/etc/init.d/bigdataHA 2014-05-06 17:57:47 UTC (rev 8208) @@ -95,7 +95,7 @@ fi fi if [ ! -f "$pidFile" ]; then - action $"`date` : `hostname` : bringing up services: " sudo -u $BD_USER -g $BD_GROUP bash -c "source /etc/default/bigdataHA ; $binDir/startHAServices" + action $"`date` : `hostname` : bringing up services: " sudo -u $BD_USER -g $BD_GROUP bash -c "source /etc/default/bigdataHA ; $binDir/startHAServices > /dev/null 2>&1 &" else echo $"`date` : `hostname` : running as $pid" fi This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-06 13:54:45
|
Revision: 8207 http://sourceforge.net/p/bigdata/code/8207 Author: thompsonbry Date: 2014-05-06 13:54:41 +0000 (Tue, 06 May 2014) Log Message: ----------- Added the jetty-jmx and jetty-jndi dependencies. The jmx dependency is necessary if you want to export the MBeans from the jetty server to another host. The jndi dependency is just foward looking - jndi provides a flexible mechanism for configuring jetty. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/.classpath branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/jetty.xml branches/BIGDATA_RELEASE_1_3_0/build.xml branches/BIGDATA_RELEASE_1_3_0/pom.xml Added Paths: ----------- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar Modified: branches/BIGDATA_RELEASE_1_3_0/.classpath =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/.classpath 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/.classpath 2014-05-06 13:54:41 UTC (rev 8207) @@ -58,6 +58,8 @@ <classpathentry exported="true" kind="lib" path="bigdata/lib/unimi/fastutil-5.1.5.jar"/> <classpathentry exported="true" kind="lib" path="bigdata/lib/lucene/lucene-analyzers-3.0.0.jar"/> <classpathentry exported="true" kind="lib" path="bigdata/lib/lucene/lucene-core-3.0.0.jar"/> + <classpathentry kind="lib" path="bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar"/> + <classpathentry kind="lib" path="bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar"/> <classpathentry exported="true" kind="con" path="org.eclipse.jdt.launching.JRE_CONTAINER"/> <classpathentry exported="true" kind="lib" path="bigdata/lib/high-scale-lib-v1.1.2.jar"/> <classpathentry exported="true" kind="lib" path="bigdata/lib/junit-ext-1.1-b3-dev.jar"/> Added: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar =================================================================== (Binary files differ) Index: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar 2014-05-06 13:54:41 UTC (rev 8207) Property changes on: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jmx-9.1.4.v20140401.jar ___________________________________________________________________ Added: svn:mime-type ## -0,0 +1 ## +application/octet-stream \ No newline at end of property Added: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar =================================================================== (Binary files differ) Index: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar 2014-05-06 13:54:41 UTC (rev 8207) Property changes on: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/jetty/jetty-jndi-9.1.4.v20140401.jar ___________________________________________________________________ Added: svn:mime-type ## -0,0 +1 ## +application/octet-stream \ No newline at end of property Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/jetty.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/jetty.xml 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/jetty.xml 2014-05-06 13:54:41 UTC (rev 8207) @@ -31,6 +31,46 @@ <Set name="detailedDump">false</Set> </Get> + <!-- =========================================================== --> + <!-- Get the platform mbean server --> + <!-- =========================================================== --> + <Call id="MBeanServer" class="java.lang.management.ManagementFactory" + name="getPlatformMBeanServer" /> + + <!-- =========================================================== --> + <!-- Initialize the Jetty MBean container --> + <!-- =========================================================== --> + <Call name="addBean"> + <Arg> + <New id="MBeanContainer" class="org.eclipse.jetty.jmx.MBeanContainer"> + <Arg> + <Ref refid="MBeanServer" /> + </Arg> + </New> + </Arg> + </Call> + + <!-- Add the static log to the MBean server. --> + <Call name="addBean"> + <Arg> + <New class="org.eclipse.jetty.util.log.Log" /> + </Arg> + </Call> + + <!-- For remote MBean access (optional) --> + <New id="ConnectorServer" class="org.eclipse.jetty.jmx.ConnectorServer"> + <Arg> + <New class="javax.management.remote.JMXServiceURL"> + <Arg type="java.lang.String">rmi</Arg> + <Arg type="java.lang.String" /> + <Arg type="java.lang.Integer"><SystemProperty name="jetty.jmxrmiport" default="1090"/></Arg> + <Arg type="java.lang.String">/jndi/rmi://<SystemProperty name="jetty.jmxrmihost" default="localhost"/>:<SystemProperty name="jetty.jmxrmiport" default="1099"/>/jmxrmi</Arg> + </New> + </Arg> + <Arg>org.eclipse.jetty.jmx:name=rmiconnectorserver</Arg> + <Call name="start" /> + </New> + <!-- =========================================================== --> <!-- Http Configuration. --> <!-- This is a common configuration instance used by all --> @@ -88,7 +128,7 @@ <Item> <!-- This is the bigdata web application. --> <New id="WebAppContext" class="org.eclipse.jetty.webapp.WebAppContext"> - <Set name="resourceBase"> + <Set name="war"> <!-- The location of the top-level of the bigdata webapp. --> <SystemProperty name="jetty.resourceBase" default="bigdata-war/src" /> </Set> @@ -98,30 +138,11 @@ <Set name="extractWAR">false</Set> </New> </Item> - <Item> - <!-- This appears to be necessary in addition to the above. --> - <!-- Without this, it will not resolve http://localhost:8080/ --> - <!-- and can fail to deliver some of the static content. --> - <New id="ResourceHandler" class="org.eclipse.jetty.server.handler.ResourceHandler"> - <Set name="resourceBase"> - <!-- The location of the top-level of the bigdata webapp. --> - <SystemProperty name="jetty.resourceBase" default="bigdata-war/src" /> - </Set> - <Set name="welcomeFiles"> - <Array type="java.lang.String"> - <Item>html/index.html</Item> - </Array> - </Set> - </New> - </Item> - <!-- <Item> - <New id="DefaultHandler" class="org.eclipse.jetty.server.handler.DefaultHandler"></New> - </Item> --> </Array> </Set> </New> </Set> - + <!-- =========================================================== --> <!-- extra server options --> <!-- =========================================================== --> Modified: branches/BIGDATA_RELEASE_1_3_0/build.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/build.xml 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/build.xml 2014-05-06 13:54:41 UTC (rev 8207) @@ -980,8 +980,12 @@ tofile="${dist.lib}/jetty-continuation.jar" /> <copy file="${bigdata-jetty.lib}/jetty-http-${jetty.version}.jar" tofile="${dist.lib}/jetty-http.jar" /> - <copy file="${bigdata-jetty.lib}/jetty-io-${jetty.version}.jar" - tofile="${dist.lib}/jetty-io.jar" /> + <copy file="${bigdata-jetty.lib}/jetty-io-${jetty.version}.jar" + tofile="${dist.lib}/jetty-io.jar" /> + <copy file="${bigdata-jetty.lib}/jetty-jmx-${jetty.version}.jar" + tofile="${dist.lib}/jetty-jmx.jar" /> + <copy file="${bigdata-jetty.lib}/jetty-jndi-${jetty.version}.jar" + tofile="${dist.lib}/jetty-jndi.jar" /> <copy file="${bigdata-jetty.lib}/jetty-server-${jetty.version}.jar" tofile="${dist.lib}/jetty-server.jar" /> <copy file="${bigdata-jetty.lib}/jetty-util-${jetty.version}.jar" @@ -1208,10 +1212,10 @@ todir="${dist.dir}/doc" /> <!-- Stage documentation from the wiki. --> - <get dest="${dist.doc}/HAJournalServer.html" + <get dest="${dist.doc}/HAJournalServer.html" ignoreerrors="true" src="http://wiki.bigdata.com/wiki/index.php/HAJournalServer?printable=yes" /> - <get dest="${dist.doc}/NanoSparqlServer.html" + <get dest="${dist.doc}/NanoSparqlServer.html" ignoreerrors="true" src="http://wiki.bigdata.com/wiki/index.php/NanoSparqlServer?printable=yes" /> @@ -1752,7 +1756,7 @@ <!-- TODO ${path.separator}${dist.lib}/bigdata-gas.jar --> <property name="javac.test.classpath" - value="${classes.dir}${path.separator}${junit.jar}${path.separator}${junit-ext.jar}${path.separator}${sesame-sparql-test.jar}${path.separator}${sesame-store-test.jar}${path.separator}${sesame-rio-test.jar}${path.separator}${dist.lib}/classserver.jar${path.separator}${dist.lib}/highscalelib.jar${path.separator}${dist.lib}/dsiutils.jar${path.separator}${dist.lib}/lgplutils.jar${path.separator}${dist.lib}/fastutil.jar${path.separator}${dist.lib}/bigdata-ganglia.jar${path.separator}${dist.lib}/icu4j.jar${path.separator}${dist.lib}/icu4j-charset.jar${path.separator}${dist.lib}/log4j.jar${path.separator}${dist.lib}/lucene-analyzer.jar${path.separator}${dist.lib}/lucene-core.jar${path.separator}${path.separator}${dist.lib}/openrdf-sesame.jar${path.separator}${dist.lib}/slf4j.jar${path.separator}${dist.lib}/jsk-lib.jar${path.separator}${dist.lib}/jsk-platform.jar${path.separator}${dist.lib}/nxparser.jar${path.separator}${dist.lib}/zookeeper.jar${path.separator}${dist.lib}/jetty-continuation.jar${path.separator}${dist.lib}/jetty-http.jar${path.separator}${dist.lib}/jetty-io.jar${path.separator}${dist.lib}/jetty-server.jar${path.separator}${dist.lib}/jetty-util.jar${path.separator}${dist.lib}/jetty-webapp.jar${path.separator}${dist.lib}/jetty-servlet.jar${path.separator}${dist.lib}/jetty-security.jar${path.separator}${dist.lib}/jetty-xml.jar${path.separator}${dist.lib}/jetty-rewrite.jar${path.separator}${dist.lib}/jetty-client.jar${path.separator}${dist.lib}/jetty-proxy.jar${path.separator}${dist.lib}/servlet-api.jar${path.separator}${dist.lib}/commons-codec.jar${path.separator}${dist.lib}/commons-fileupload.jar${path.separator}${dist.lib}/commons-io.jar${path.separator}${dist.lib}/commons-logging.jar${path.separator}${dist.lib}/httpclient.jar${path.separator}${dist.lib}/httpclient-cache.jar${path.separator}${dist.lib}/httpcore.jar${path.separator}${dist.lib}/httpmime.jar" /> + value="${classes.dir}${path.separator}${junit.jar}${path.separator}${junit-ext.jar}${path.separator}${sesame-sparql-test.jar}${path.separator}${sesame-store-test.jar}${path.separator}${sesame-rio-test.jar}${path.separator}${dist.lib}/classserver.jar${path.separator}${dist.lib}/highscalelib.jar${path.separator}${dist.lib}/dsiutils.jar${path.separator}${dist.lib}/lgplutils.jar${path.separator}${dist.lib}/fastutil.jar${path.separator}${dist.lib}/bigdata-ganglia.jar${path.separator}${dist.lib}/icu4j.jar${path.separator}${dist.lib}/icu4j-charset.jar${path.separator}${dist.lib}/log4j.jar${path.separator}${dist.lib}/lucene-analyzer.jar${path.separator}${dist.lib}/lucene-core.jar${path.separator}${path.separator}${dist.lib}/openrdf-sesame.jar${path.separator}${dist.lib}/slf4j.jar${path.separator}${dist.lib}/jsk-lib.jar${path.separator}${dist.lib}/jsk-platform.jar${path.separator}${dist.lib}/nxparser.jar${path.separator}${dist.lib}/zookeeper.jar${path.separator}${dist.lib}/jetty-continuation.jar${path.separator}${dist.lib}/jetty-http.jar${path.separator}${dist.lib}/jetty-io.jar${path.separator}${dist.lib}/jetty-jmx.jar${path.separator}${dist.lib}/jetty-jndi.jar${path.separator}${dist.lib}/jetty-server.jar${path.separator}${dist.lib}/jetty-util.jar${path.separator}${dist.lib}/jetty-webapp.jar${path.separator}${dist.lib}/jetty-servlet.jar${path.separator}${dist.lib}/jetty-security.jar${path.separator}${dist.lib}/jetty-xml.jar${path.separator}${dist.lib}/jetty-rewrite.jar${path.separator}${dist.lib}/jetty-client.jar${path.separator}${dist.lib}/jetty-proxy.jar${path.separator}${dist.lib}/servlet-api.jar${path.separator}${dist.lib}/commons-codec.jar${path.separator}${dist.lib}/commons-fileupload.jar${path.separator}${dist.lib}/commons-io.jar${path.separator}${dist.lib}/commons-logging.jar${path.separator}${dist.lib}/httpclient.jar${path.separator}${dist.lib}/httpclient-cache.jar${path.separator}${dist.lib}/httpcore.jar${path.separator}${dist.lib}/httpmime.jar" /> <echo>javac </echo> @@ -2112,6 +2116,8 @@ <pathelement location="${dist.lib}/jetty-continuation.jar" /> <pathelement location="${dist.lib}/jetty-http.jar" /> <pathelement location="${dist.lib}/jetty-io.jar" /> + <pathelement location="${dist.lib}/jetty-jmx.jar" /> + <pathelement location="${dist.lib}/jetty-jndi.jar" /> <pathelement location="${dist.lib}/jetty-server.jar" /> <pathelement location="${dist.lib}/jetty-util.jar" /> <pathelement location="${dist.lib}/jetty-webapp.jar" /> @@ -2133,7 +2139,7 @@ </path> <property name="run.class.path" - value="${junit.jar}${path.separator}${bigdata-test.jar}${path.separator}${junit-ext.jar}${path.separator}${sesame-sparql-test.jar}${path.separator}${sesame-store-test.jar}${path.separator}${sesame-rio-test.jar}${path.separator}${dist.lib}/bigdata.jar${path.separator}${dist.lib}/colt.jar${path.separator}${dist.lib}/highscalelib.jar${path.separator}${dist.lib}/dsiutils.jar${path.separator}${dist.lib}/lgplutils.jar${path.separator}${dist.lib}/fastutil.jar${path.separator}${dist.lib}/bigdata-ganglia.jar${path.separator}${dist.lib}/bigdata-gas${path.separator}${dist.lib}/icu4j.jar${path.separator}${dist.lib}/icu4j-charset.jar${path.separator}${dist.lib}/jsk-lib.jar${path.separator}${dist.lib}/jsk-platform.jar${path.separator}${dist.lib}/log4j.jar${path.separator}${dist.lib}/lucene-analyzer.jar${path.separator}${dist.lib}/lucene-core.jar${path.separator}${dist.lib}/openrdf-sesame.jar${path.separator}${dist.lib}/slf4j.jar${path.separator}${dist.lib}/slf4j-log4j.jar${path.separator}${dist.lib}/nxparser.jar${path.separator}${dist.lib}/zookeeper.jar${path.separator}${dist.lib}/jetty-continuation.jar${path.separator}${dist.lib}/jetty-http.jar${path.separator}${dist.lib}/jetty-io.jar${path.separator}${dist.lib}/jetty-server.jar${path.separator}${dist.lib}/jetty-util.jar${path.separator}${dist.lib}/jetty-webapp.jar${path.separator}${dist.lib}/jetty-servlet.jar${path.separator}${dist.lib}/jetty-security.jar${path.separator}${dist.lib}/jetty-xml.jar${path.separator}${dist.lib}/jetty-rewrite.jar${path.separator}${dist.lib}/jetty-client.jar${path.separator}${dist.lib}/jetty-proxy.jar${path.separator}${dist.lib}/servlet-api.jar${path.separator}${dist.lib}/commons-codec.jar${path.separator}${dist.lib}/commons-fileupload.jar${path.separator}${dist.lib}/commons-io.jar${path.separator}${dist.lib}/commons-logging.jar${path.separator}${dist.lib}/httpclient.jar${path.separator}${dist.lib}/httpclient-cache.jar${path.separator}${dist.lib}/httpcore.jar${path.separator}${dist.lib}/httpmime.jar" /> + value="${junit.jar}${path.separator}${bigdata-test.jar}${path.separator}${junit-ext.jar}${path.separator}${sesame-sparql-test.jar}${path.separator}${sesame-store-test.jar}${path.separator}${sesame-rio-test.jar}${path.separator}${dist.lib}/bigdata.jar${path.separator}${dist.lib}/colt.jar${path.separator}${dist.lib}/highscalelib.jar${path.separator}${dist.lib}/dsiutils.jar${path.separator}${dist.lib}/lgplutils.jar${path.separator}${dist.lib}/fastutil.jar${path.separator}${dist.lib}/bigdata-ganglia.jar${path.separator}${dist.lib}/bigdata-gas${path.separator}${dist.lib}/icu4j.jar${path.separator}${dist.lib}/icu4j-charset.jar${path.separator}${dist.lib}/jsk-lib.jar${path.separator}${dist.lib}/jsk-platform.jar${path.separator}${dist.lib}/log4j.jar${path.separator}${dist.lib}/lucene-analyzer.jar${path.separator}${dist.lib}/lucene-core.jar${path.separator}${dist.lib}/openrdf-sesame.jar${path.separator}${dist.lib}/slf4j.jar${path.separator}${dist.lib}/slf4j-log4j.jar${path.separator}${dist.lib}/nxparser.jar${path.separator}${dist.lib}/zookeeper.jar${path.separator}${dist.lib}/jetty-continuation.jar${path.separator}${dist.lib}/jetty-http.jar${path.separator}${dist.lib}/jetty-io.jar${path.separator}${dist.lib}/jetty-jmx.jar${path.separator}${dist.lib}/jetty-jndi.jar${path.separator}${dist.lib}/jetty-server.jar${path.separator}${dist.lib}/jetty-util.jar${path.separator}${dist.lib}/jetty-webapp.jar${path.separator}${dist.lib}/jetty-servlet.jar${path.separator}${dist.lib}/jetty-security.jar${path.separator}${dist.lib}/jetty-xml.jar${path.separator}${dist.lib}/jetty-rewrite.jar${path.separator}${dist.lib}/jetty-client.jar${path.separator}${dist.lib}/jetty-proxy.jar${path.separator}${dist.lib}/servlet-api.jar${path.separator}${dist.lib}/commons-codec.jar${path.separator}${dist.lib}/commons-fileupload.jar${path.separator}${dist.lib}/commons-io.jar${path.separator}${dist.lib}/commons-logging.jar${path.separator}${dist.lib}/httpclient.jar${path.separator}${dist.lib}/httpclient-cache.jar${path.separator}${dist.lib}/httpcore.jar${path.separator}${dist.lib}/httpmime.jar" /> <echo> classpath: ${run.class.path} </echo> @@ -2279,6 +2285,8 @@ <sysproperty key="jetty-continuation.jar" value="${dist.lib}/jetty-continuation.jar" /> <sysproperty key="jetty-http.jar" value="${dist.lib}/jetty-http.jar" /> <sysproperty key="jetty-io.jar" value="${dist.lib}/jetty-io.jar" /> + <sysproperty key="jetty-jmx.jar" value="${dist.lib}/jetty-jmx.jar" /> + <sysproperty key="jetty-jndi.jar" value="${dist.lib}/jetty-jndi.jar" /> <sysproperty key="jetty-server.jar" value="${dist.lib}/jetty-server.jar" /> <sysproperty key="jetty-util.jar" value="${dist.lib}/jetty-util.jar" /> <sysproperty key="jetty-webapp.jar" value="${dist.lib}/jetty-webapp.jar" /> Modified: branches/BIGDATA_RELEASE_1_3_0/pom.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/pom.xml 2014-05-06 10:49:40 UTC (rev 8206) +++ branches/BIGDATA_RELEASE_1_3_0/pom.xml 2014-05-06 13:54:41 UTC (rev 8207) @@ -409,6 +409,8 @@ jetty-continuation jetty-http jetty-io + jetty-jmx + jetty-jndi jetty-server jetty-util jetty-webapp @@ -435,6 +437,16 @@ <artifactId>jetty-io</artifactId> <version>${jetty.version}</version> </dependency> + <dependency> + <groupId>org.eclipse.jetty</groupId> + <artifactId>jetty-jmx</artifactId> + <version>${jetty.version}</version> + </dependency> + <dependency> + <groupId>org.eclipse.jetty</groupId> + <artifactId>jetty-jndi</artifactId> + <version>${jetty.version}</version> + </dependency> <dependency> <groupId>org.eclipse.jetty</groupId> <artifactId>jetty-server</artifactId> This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-06 10:49:45
|
Revision: 8206 http://sourceforge.net/p/bigdata/code/8206 Author: thompsonbry Date: 2014-05-06 10:49:40 +0000 (Tue, 06 May 2014) Log Message: ----------- Syntax error in the HAConfig file. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-06 10:15:56 UTC (rev 8205) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-06 10:49:40 UTC (rev 8206) @@ -63,7 +63,7 @@ private static fedname = ConfigMath.getProperty("FEDNAME","benchmark"); // The RMI port for the HAGlue interface (may be ZERO for a random port). - private static rmiPort = Integer.parseInt(ConfigMath.getProperty("RMI_PORT","9080"));; + private static rmiPort = Integer.parseInt(ConfigMath.getProperty("RMI_PORT","9080")); // write replication pipeline port (listener). private static haPort = Integer.parseInt(ConfigMath.getProperty("HA_PORT","9090")); This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-06 10:16:00
|
Revision: 8205 http://sourceforge.net/p/bigdata/code/8205 Author: thompsonbry Date: 2014-05-06 10:15:56 +0000 (Tue, 06 May 2014) Log Message: ----------- Added an RMI_PORT for the exporter for the HAGlue interface. This can be set from the startHAServices script. The ganglia policy needs to chose a target host in inverse proportion to the workload on that host. See #624 (HA LBS) Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-A.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-B.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-C.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-A.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-B.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-C.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-D.config branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-E.config branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices branches/BIGDATA_RELEASE_1_3_0/src/resources/etc/default/bigdataHA Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-A.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-A.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-A.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,6 +60,9 @@ private static fedname = "benchmark"; + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = 9080; + // write replication pipeline port (listener). private static haPort = 9090; @@ -250,6 +253,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + } /* Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-B.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-B.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-B.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,9 +60,9 @@ private static fedname = "benchmark"; - // NanoSparqlServer (http) port. - private static nssPort = ConfigMath.add(8090,1); - + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,1); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,1); @@ -252,6 +252,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + } /* Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-C.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-C.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournal-C.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,6 +60,9 @@ private static fedname = "benchmark"; + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,1); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,2); @@ -249,6 +252,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + } /* Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-A.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-A.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-A.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,6 +60,9 @@ private static fedname = "benchmark"; + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = 9080; + // write replication pipeline port (listener). private static haPort = 9090; @@ -257,6 +260,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + // Use the overridden version of the HAJournal by default so we get the // HAGlueTest API for every test. HAJournalClass = "com.bigdata.journal.jini.ha.HAJournalTest"; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-B.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-B.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-B.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,6 +60,9 @@ private static fedname = "benchmark"; + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,1); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,1); @@ -256,6 +259,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + // Use the overridden version of the HAJournal by default so we get the // HAGlueTest API for every test. HAJournalClass = "com.bigdata.journal.jini.ha.HAJournalTest"; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-C.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-C.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-C.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,6 +60,9 @@ private static fedname = "benchmark"; + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,2); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,2); @@ -256,6 +259,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + // Use the overridden version of the HAJournal by default so we get the // HAGlueTest API for every test. HAJournalClass = "com.bigdata.journal.jini.ha.HAJournalTest"; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-D.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-D.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-D.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,9 +60,9 @@ private static fedname = "benchmark"; - // NanoSparqlServer (http) port. - private static nssPort = ConfigMath.add(8090,3); - + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,3); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,3); @@ -259,6 +259,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + // Use the overridden version of the HAJournal by default so we get the // HAGlueTest API for every test. HAJournalClass = "com.bigdata.journal.jini.ha.HAJournalTest"; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-E.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-E.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournal-E.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -60,9 +60,9 @@ private static fedname = "benchmark"; - // NanoSparqlServer (http) port. - private static nssPort = ConfigMath.add(8090,4); - + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = ConfigMath.add(9080,4); + // write replication pipeline port (listener). private static haPort = ConfigMath.add(9090,4); @@ -259,6 +259,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + // Use the overridden version of the HAJournal by default so we get the // HAGlueTest API for every test. HAJournalClass = "com.bigdata.journal.jini.ha.HAJournalTest"; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-06 10:15:56 UTC (rev 8205) @@ -717,7 +717,12 @@ * Stochastically select the target host based on the current host * workload. * - * We need to ignore any host that is not joined with the met quorum.... + * Note: We need to ignore any host that is not joined with the met + * quorum.... + * + * Note: The host is selected with a probability that is INVERSELY + * proportional to normalized host load. If the normalized host load is + * .75, then the host is selected with a probability of .25. */ final double d = rand.nextDouble(); @@ -726,13 +731,13 @@ final List<ServiceScore> foundServices = new LinkedList<ServiceScore>(); for (int i = 0; i < hostScores.length; i++) { hostScore = hostScores[i]; - sum += hostScore.score; + sum += 1d - hostScore.score; // Note: Choice is inversely proportional to normalized workload! if (hostScore.hostname == null) // can't use w/o hostname. continue; if (d > sum) // scan further. continue; /* - * We found a host having a position in the cumultive ordering of + * We found a host having a position in the cumulative ordering of * the normalized host workloads that is GTE to the random number. * Now we need to find one (or more) service(s) on that host. * Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-06 10:15:56 UTC (rev 8205) @@ -62,6 +62,9 @@ */ private static fedname = ConfigMath.getProperty("FEDNAME","benchmark"); + // The RMI port for the HAGlue interface (may be ZERO for a random port). + private static rmiPort = Integer.parseInt(ConfigMath.getProperty("RMI_PORT","9080"));; + // write replication pipeline port (listener). private static haPort = Integer.parseInt(ConfigMath.getProperty("HA_PORT","9090")); @@ -304,6 +307,9 @@ replicationFactor = bigdata.replicationFactor; + exporter = new BasicJeriExporter(TcpServerEndpoint.getInstance(bigdata.rmiPort), + new BasicILFactory()); + haLogDir = bigdata.haLogDir; snapshotDir = bigdata.snapshotDir; Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices 2014-05-06 10:15:56 UTC (rev 8205) @@ -80,6 +80,7 @@ -DGROUPS=${GROUPS}\ -DLOCATORS=${LOCATORS}\ -DZK_SERVERS=${ZK_SERVERS}\ + -DRMI_PORT=${RMI_PORT}\ -DHA_PORT=${HA_PORT}\ "-Dcom.bigdata.hostname=${BIGDATA_HOSTNAME}"\ "-Djetty.port=${JETTY_PORT}"\ Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/etc/default/bigdataHA =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/etc/default/bigdataHA 2014-05-06 00:30:25 UTC (rev 8204) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/etc/default/bigdataHA 2014-05-06 10:15:56 UTC (rev 8205) @@ -38,6 +38,7 @@ # All of these have defaults. Override as necessary. #export REPLICATION_FACTOR=3 +#export RMI_PORT=9080 #export HA_PORT=9090 #export JETTY_PORT=8080 #export JETTY_XML=var/jetty/WEB-INF/jetty.xml This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tob...@us...> - 2014-05-06 00:30:31
|
Revision: 8204 http://sourceforge.net/p/bigdata/code/8204 Author: tobycraig Date: 2014-05-06 00:30:25 +0000 (Tue, 06 May 2014) Log Message: ----------- Advanced commands (monitor and analytic) enabled for update Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/css/style.css branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/index.html branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/css/style.css =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/css/style.css 2014-05-05 19:47:44 UTC (rev 8203) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/css/style.css 2014-05-06 00:30:25 UTC (rev 8204) @@ -250,7 +250,7 @@ border: none; } -#advanced-features, #query-response, #query-pagination, #query-explanation, #query-export-container, #update-response, #update-clear-container, #explore-results, #namespace-properties { +.advanced-features, #query-response, #query-pagination, #query-explanation, #query-export-container, #update-response, #update-clear-container, #explore-results, #namespace-properties { display: none; } @@ -353,3 +353,10 @@ text-align: right; } +iframe { + display: none; + width: 100%; + height: 300px; + border: 1px solid #e1e1e1; + box-sizing: border-box; +} Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/index.html =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/index.html 2014-05-05 19:47:44 UTC (rev 8203) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/index.html 2014-05-06 00:30:25 UTC (rev 8204) @@ -50,7 +50,7 @@ <option value="path">File Path or URL</option> </select> <label for="rdf-type">Format:</label> - <select id="rdf-type"> + <select id="rdf-type" disabled> <option value="n-quads">N-Quads</option> <option value="n-triples">N-Triples</option> <option value="n3">Notation3</option> @@ -61,6 +61,11 @@ <option value="turtle">Turtle</option> </select> </p> + <a href="#" class="advanced-features-toggle">Advanced features</a> + <div class="advanced-features"> + <input type="checkbox" id="update-analytic"> <label for="update-analytic">Analytic</label> + <input type="checkbox" id="update-monitor"> <label for="update-monitor">Monitor</label> + </div> <hr class="shadow"> <button id="update-update">Update</button> @@ -68,6 +73,7 @@ <div class="box" id="update-response"> <pre></pre> + <iframe name="update-response-container"></iframe> </div> <div class="box" id="update-clear-container"> @@ -86,9 +92,9 @@ <div id="query-errors"></div> <textarea id="query-box" name="query" placeholder="(Input a SPARQL query)"></textarea> - <a href="#" id="advanced-features-toggle">Advanced features</a> + <a href="#" class="advanced-features-toggle">Advanced features</a> - <div id="advanced-features"> + <div class="advanced-features"> <input type="checkbox" id="query-explain"> <label for="query-explain">Explain</label> <input type="checkbox" id="query-details"> <label for="query-explain">(Details)</label> <input type="checkbox" name="analytic" value="true" id="query-analytic"> <label for="query-analytic">Analytic</label> Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-05 19:47:44 UTC (rev 8203) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-06 00:30:25 UTC (rev 8204) @@ -340,8 +340,16 @@ if(type == 'rdf') { $('#rdf-type').val(format); } + showUpdateOptions(type); } +$('#update-type').change(function() { showUpdateOptions(this.value) }); + +function showUpdateOptions(type) { + $('#rdf-type, label[for="rdf-type"]').attr('disabled', type != 'rdf'); + $('#update-tab .advanced-features input').attr('disabled', type != 'sparql'); +} + // .xml is used for both RDF and TriX, assume it's RDF // We could check the parent element to see which it is var rdf_types = {'nq': 'n-quads', @@ -379,8 +387,14 @@ $('#update-update').click(submitUpdate); function submitUpdate(e) { + // Updates are submitted as a regular form for SPARQL updates in monitor mode, and via AJAX for non-monitor SPARQL, RDF & file path updates. + // When submitted as a regular form, the output is sent to an iframe. This is to allow monitor mode to work. + // jQuery only gives us data when the request is complete, so we wouldn't see monitor results as they come in. + e.preventDefault(); + $('#update-response').show(); + var settings = { type: 'POST', data: FILE_CONTENTS == null ? $('#update-box').val() : FILE_CONTENTS, @@ -391,7 +405,25 @@ // determine action based on type switch($('#update-type').val()) { case 'sparql': + // see if monitor mode is on + if($('#update-monitor').is(':checked')) { + // create form and submit it, sending output to the iframe + var form = $('<form method="POST" target="update-response-container">') + .attr('action', NAMESPACE_URL) + .append($('<input name="update">').val(settings.data)) + .append('<input name="monitor" value="true">'); + if($('#update-analytic').is(':checked')) { + form.append('<input name="analytic" value="true">') + } + form.submit(); + $('#update-response iframe, #update-clear-container').show(); + $('#update-response pre').hide(); + return; + } settings.data = 'update=' + encodeURIComponent(settings.data); + if($('#update-analytic').is(':checked')) { + settings.data += '&analytic=true'; + } settings.success = updateResponseHTML; break; case 'rdf': @@ -411,36 +443,39 @@ break; } - $('#update-response').show(); - $('#update-response pre').html('Data loading...'); + $('#update-response pre').show().html('Data loading...'); $.ajax(NAMESPACE_URL, settings); } $('#update-clear').click(function() { - $('#update-response, #update-clear').hide(); + $('#update-response, #update-clear-container').hide(); $('#update-response pre').text(''); + $('#update-response iframe').attr('src', 'about:blank'); }); -$('#advanced-features-toggle').click(function() { - $('#advanced-features').toggle(); +$('.advanced-features-toggle').click(function() { + $(this).next('.advanced-features').toggle(); return false; }); function updateResponseHTML(data) { - $('#update-response, #update-clear').show(); + $('#update-response, #update-clear-container').show(); + $('#update-response iframe').attr('src', 'about:blank').hide(); $('#update-response pre').html(data); } function updateResponseXML(data) { var modified = data.childNodes[0].attributes['modified'].value; var milliseconds = data.childNodes[0].attributes['milliseconds'].value; - $('#update-response, #update-clear').show(); + $('#update-response, #update-clear-container').show(); + $('#update-response iframe').attr('src', 'about:blank').hide(); $('#update-response pre').text('Modified: ' + modified + '\nMilliseconds: ' + milliseconds); } function updateResponseError(jqXHR, textStatus, errorThrown) { - $('#update-response, #update-clear').show(); + $('#update-response, #update-clear-container').show(); + $('#update-response iframe').attr('src', 'about:blank').hide(); $('#update-response pre').text('Error! ' + textStatus + ' ' + jqXHR.statusText); highlightError(jqXHR.statusText, 'update'); } This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-05 19:47:49
|
Revision: 8203 http://sourceforge.net/p/bigdata/code/8203 Author: thompsonbry Date: 2014-05-05 19:47:44 +0000 (Mon, 05 May 2014) Log Message: ----------- Replaced the old trac and wiki links in the release notes with the new trac and wiki links. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt 2014-05-05 19:01:03 UTC (rev 8202) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt 2014-05-05 19:47:44 UTC (rev 8203) @@ -58,332 +58,332 @@ 1.3.0: -- http://sourceforge.net/apps/trac/bigdata/ticket/530 (Journal HA) -- http://sourceforge.net/apps/trac/bigdata/ticket/621 (Coalesce write cache records and install reads in cache) -- http://sourceforge.net/apps/trac/bigdata/ticket/623 (HA TXS) -- http://sourceforge.net/apps/trac/bigdata/ticket/639 (Remove triple-buffering in RWStore) -- http://sourceforge.net/apps/trac/bigdata/ticket/645 (HA backup) -- http://sourceforge.net/apps/trac/bigdata/ticket/646 (River not compatible with newer 1.6.0 and 1.7.0 JVMs) -- http://sourceforge.net/apps/trac/bigdata/ticket/648 (Add a custom function to use full text index for filtering.) -- http://sourceforge.net/apps/trac/bigdata/ticket/651 (RWS test failure) -- http://sourceforge.net/apps/trac/bigdata/ticket/652 (Compress write cache blocks for replication and in HALogs) -- http://sourceforge.net/apps/trac/bigdata/ticket/662 (Latency on followers during commit on leader) -- http://sourceforge.net/apps/trac/bigdata/ticket/663 (Issue with OPTIONAL blocks) -- http://sourceforge.net/apps/trac/bigdata/ticket/664 (RWStore needs post-commit protocol) -- http://sourceforge.net/apps/trac/bigdata/ticket/665 (HA3 LOAD non-responsive with node failure) -- http://sourceforge.net/apps/trac/bigdata/ticket/666 (Occasional CI deadlock in HALogWriter testConcurrentRWWriterReader) -- http://sourceforge.net/apps/trac/bigdata/ticket/670 (Accumulating HALog files cause latency for HA commit) -- http://sourceforge.net/apps/trac/bigdata/ticket/671 (Query on follower fails during UPDATE on leader) -- http://sourceforge.net/apps/trac/bigdata/ticket/673 (DGC in release time consensus protocol causes native thread leak in HAJournalServer at each commit) -- http://sourceforge.net/apps/trac/bigdata/ticket/674 (WCS write cache compaction causes errors in RWS postHACommit()) -- http://sourceforge.net/apps/trac/bigdata/ticket/676 (Bad patterns for timeout computations) -- http://sourceforge.net/apps/trac/bigdata/ticket/677 (HA deadlock under UPDATE + QUERY) -- http://sourceforge.net/apps/trac/bigdata/ticket/678 (DGC Thread and Open File Leaks: sendHALogForWriteSet()) -- http://sourceforge.net/apps/trac/bigdata/ticket/679 (HAJournalServer can not restart due to logically empty log file) -- http://sourceforge.net/apps/trac/bigdata/ticket/681 (HAJournalServer deadlock: pipelineRemove() and getLeaderId()) -- http://sourceforge.net/apps/trac/bigdata/ticket/684 (Optimization with skos altLabel) -- http://sourceforge.net/apps/trac/bigdata/ticket/686 (Consensus protocol does not detect clock skew correctly) -- http://sourceforge.net/apps/trac/bigdata/ticket/687 (HAJournalServer Cache not populated) -- http://sourceforge.net/apps/trac/bigdata/ticket/689 (Missing URL encoding in RemoteRepositoryManager) -- http://sourceforge.net/apps/trac/bigdata/ticket/690 (Error when using the alias "a" instead of rdf:type for a multipart insert) -- http://sourceforge.net/apps/trac/bigdata/ticket/691 (Failed to re-interrupt thread in HAJournalServer) -- http://sourceforge.net/apps/trac/bigdata/ticket/692 (Failed to re-interrupt thread) -- http://sourceforge.net/apps/trac/bigdata/ticket/693 (OneOrMorePath SPARQL property path expression ignored) -- http://sourceforge.net/apps/trac/bigdata/ticket/694 (Transparently cancel update/query in RemoteRepository) -- http://sourceforge.net/apps/trac/bigdata/ticket/695 (HAJournalServer reports "follower" but is in SeekConsensus and is not participating in commits.) -- http://sourceforge.net/apps/trac/bigdata/ticket/701 (Problems in BackgroundTupleResult) -- http://sourceforge.net/apps/trac/bigdata/ticket/702 (InvocationTargetException on /namespace call) -- http://sourceforge.net/apps/trac/bigdata/ticket/704 (ask does not return json) -- http://sourceforge.net/apps/trac/bigdata/ticket/705 (Race between QueryEngine.putIfAbsent() and shutdownNow()) -- http://sourceforge.net/apps/trac/bigdata/ticket/706 (MultiSourceSequentialCloseableIterator.nextSource() can throw NPE) -- http://sourceforge.net/apps/trac/bigdata/ticket/707 (BlockingBuffer.close() does not unblock threads) -- http://sourceforge.net/apps/trac/bigdata/ticket/708 (BIND heisenbug - race condition on select query with BIND) -- http://sourceforge.net/apps/trac/bigdata/ticket/711 (sparql protocol: mime type application/sparql-query) -- http://sourceforge.net/apps/trac/bigdata/ticket/712 (SELECT ?x { OPTIONAL { ?x eg:doesNotExist eg:doesNotExist } } incorrect) -- http://sourceforge.net/apps/trac/bigdata/ticket/715 (Interrupt of thread submitting a query for evaluation does not always terminate the AbstractRunningQuery) -- http://sourceforge.net/apps/trac/bigdata/ticket/716 (Verify that IRunningQuery instances (and nested queries) are correctly cancelled when interrupted) -- http://sourceforge.net/apps/trac/bigdata/ticket/718 (HAJournalServer needs to handle ZK client connection loss) -- http://sourceforge.net/apps/trac/bigdata/ticket/720 (HA3 simultaneous service start failure) -- http://sourceforge.net/apps/trac/bigdata/ticket/723 (HA asynchronous tasks must be canceled when invariants are changed) -- http://sourceforge.net/apps/trac/bigdata/ticket/725 (FILTER EXISTS in subselect) -- http://sourceforge.net/apps/trac/bigdata/ticket/726 (Logically empty HALog for committed transaction) -- http://sourceforge.net/apps/trac/bigdata/ticket/727 (DELETE/INSERT fails with OPTIONAL non-matching WHERE) -- http://sourceforge.net/apps/trac/bigdata/ticket/728 (Refactor to create HAClient) -- http://sourceforge.net/apps/trac/bigdata/ticket/729 (ant bundleJar not working) -- http://sourceforge.net/apps/trac/bigdata/ticket/731 (CBD and Update leads to 500 status code) -- http://sourceforge.net/apps/trac/bigdata/ticket/732 (describe statement limit does not work) -- http://sourceforge.net/apps/trac/bigdata/ticket/733 (Range optimizer not optimizing Slice service) -- http://sourceforge.net/apps/trac/bigdata/ticket/734 (two property paths interfere) -- http://sourceforge.net/apps/trac/bigdata/ticket/736 (MIN() malfunction) -- http://sourceforge.net/apps/trac/bigdata/ticket/737 (class cast exception) -- http://sourceforge.net/apps/trac/bigdata/ticket/739 (Inconsistent treatment of bind and optional property path) -- http://sourceforge.net/apps/trac/bigdata/ticket/741 (ctc-striterators should build as independent top-level project (Apache2)) -- http://sourceforge.net/apps/trac/bigdata/ticket/743 (AbstractTripleStore.destroy() does not filter for correct prefix) -- http://sourceforge.net/apps/trac/bigdata/ticket/746 (Assertion error) -- http://sourceforge.net/apps/trac/bigdata/ticket/747 (BOUND bug) -- http://sourceforge.net/apps/trac/bigdata/ticket/748 (incorrect join with subselect renaming vars) -- http://sourceforge.net/apps/trac/bigdata/ticket/754 (Failure to setup SERVICE hook and changeLog for Unisolated and Read/Write connections) -- http://sourceforge.net/apps/trac/bigdata/ticket/755 (Concurrent QuorumActors can interfere leading to failure to progress) -- http://sourceforge.net/apps/trac/bigdata/ticket/756 (order by and group_concat) -- http://sourceforge.net/apps/trac/bigdata/ticket/760 (Code review on 2-phase commit protocol) -- http://sourceforge.net/apps/trac/bigdata/ticket/764 (RESYNC failure (HA)) -- http://sourceforge.net/apps/trac/bigdata/ticket/770 (alpp ordering) -- http://sourceforge.net/apps/trac/bigdata/ticket/772 (Query timeout only checked at operator start/stop.) -- http://sourceforge.net/apps/trac/bigdata/ticket/776 (Closed as duplicate of #490) -- http://sourceforge.net/apps/trac/bigdata/ticket/778 (HA Leader fail results in transient problem with allocations on other services) -- http://sourceforge.net/apps/trac/bigdata/ticket/783 (Operator Alerts (HA)) +- http://trac.bigdata.com/ticket/530 (Journal HA) +- http://trac.bigdata.com/ticket/621 (Coalesce write cache records and install reads in cache) +- http://trac.bigdata.com/ticket/623 (HA TXS) +- http://trac.bigdata.com/ticket/639 (Remove triple-buffering in RWStore) +- http://trac.bigdata.com/ticket/645 (HA backup) +- http://trac.bigdata.com/ticket/646 (River not compatible with newer 1.6.0 and 1.7.0 JVMs) +- http://trac.bigdata.com/ticket/648 (Add a custom function to use full text index for filtering.) +- http://trac.bigdata.com/ticket/651 (RWS test failure) +- http://trac.bigdata.com/ticket/652 (Compress write cache blocks for replication and in HALogs) +- http://trac.bigdata.com/ticket/662 (Latency on followers during commit on leader) +- http://trac.bigdata.com/ticket/663 (Issue with OPTIONAL blocks) +- http://trac.bigdata.com/ticket/664 (RWStore needs post-commit protocol) +- http://trac.bigdata.com/ticket/665 (HA3 LOAD non-responsive with node failure) +- http://trac.bigdata.com/ticket/666 (Occasional CI deadlock in HALogWriter testConcurrentRWWriterReader) +- http://trac.bigdata.com/ticket/670 (Accumulating HALog files cause latency for HA commit) +- http://trac.bigdata.com/ticket/671 (Query on follower fails during UPDATE on leader) +- http://trac.bigdata.com/ticket/673 (DGC in release time consensus protocol causes native thread leak in HAJournalServer at each commit) +- http://trac.bigdata.com/ticket/674 (WCS write cache compaction causes errors in RWS postHACommit()) +- http://trac.bigdata.com/ticket/676 (Bad patterns for timeout computations) +- http://trac.bigdata.com/ticket/677 (HA deadlock under UPDATE + QUERY) +- http://trac.bigdata.com/ticket/678 (DGC Thread and Open File Leaks: sendHALogForWriteSet()) +- http://trac.bigdata.com/ticket/679 (HAJournalServer can not restart due to logically empty log file) +- http://trac.bigdata.com/ticket/681 (HAJournalServer deadlock: pipelineRemove() and getLeaderId()) +- http://trac.bigdata.com/ticket/684 (Optimization with skos altLabel) +- http://trac.bigdata.com/ticket/686 (Consensus protocol does not detect clock skew correctly) +- http://trac.bigdata.com/ticket/687 (HAJournalServer Cache not populated) +- http://trac.bigdata.com/ticket/689 (Missing URL encoding in RemoteRepositoryManager) +- http://trac.bigdata.com/ticket/690 (Error when using the alias "a" instead of rdf:type for a multipart insert) +- http://trac.bigdata.com/ticket/691 (Failed to re-interrupt thread in HAJournalServer) +- http://trac.bigdata.com/ticket/692 (Failed to re-interrupt thread) +- http://trac.bigdata.com/ticket/693 (OneOrMorePath SPARQL property path expression ignored) +- http://trac.bigdata.com/ticket/694 (Transparently cancel update/query in RemoteRepository) +- http://trac.bigdata.com/ticket/695 (HAJournalServer reports "follower" but is in SeekConsensus and is not participating in commits.) +- http://trac.bigdata.com/ticket/701 (Problems in BackgroundTupleResult) +- http://trac.bigdata.com/ticket/702 (InvocationTargetException on /namespace call) +- http://trac.bigdata.com/ticket/704 (ask does not return json) +- http://trac.bigdata.com/ticket/705 (Race between QueryEngine.putIfAbsent() and shutdownNow()) +- http://trac.bigdata.com/ticket/706 (MultiSourceSequentialCloseableIterator.nextSource() can throw NPE) +- http://trac.bigdata.com/ticket/707 (BlockingBuffer.close() does not unblock threads) +- http://trac.bigdata.com/ticket/708 (BIND heisenbug - race condition on select query with BIND) +- http://trac.bigdata.com/ticket/711 (sparql protocol: mime type application/sparql-query) +- http://trac.bigdata.com/ticket/712 (SELECT ?x { OPTIONAL { ?x eg:doesNotExist eg:doesNotExist } } incorrect) +- http://trac.bigdata.com/ticket/715 (Interrupt of thread submitting a query for evaluation does not always terminate the AbstractRunningQuery) +- http://trac.bigdata.com/ticket/716 (Verify that IRunningQuery instances (and nested queries) are correctly cancelled when interrupted) +- http://trac.bigdata.com/ticket/718 (HAJournalServer needs to handle ZK client connection loss) +- http://trac.bigdata.com/ticket/720 (HA3 simultaneous service start failure) +- http://trac.bigdata.com/ticket/723 (HA asynchronous tasks must be canceled when invariants are changed) +- http://trac.bigdata.com/ticket/725 (FILTER EXISTS in subselect) +- http://trac.bigdata.com/ticket/726 (Logically empty HALog for committed transaction) +- http://trac.bigdata.com/ticket/727 (DELETE/INSERT fails with OPTIONAL non-matching WHERE) +- http://trac.bigdata.com/ticket/728 (Refactor to create HAClient) +- http://trac.bigdata.com/ticket/729 (ant bundleJar not working) +- http://trac.bigdata.com/ticket/731 (CBD and Update leads to 500 status code) +- http://trac.bigdata.com/ticket/732 (describe statement limit does not work) +- http://trac.bigdata.com/ticket/733 (Range optimizer not optimizing Slice service) +- http://trac.bigdata.com/ticket/734 (two property paths interfere) +- http://trac.bigdata.com/ticket/736 (MIN() malfunction) +- http://trac.bigdata.com/ticket/737 (class cast exception) +- http://trac.bigdata.com/ticket/739 (Inconsistent treatment of bind and optional property path) +- http://trac.bigdata.com/ticket/741 (ctc-striterators should build as independent top-level project (Apache2)) +- http://trac.bigdata.com/ticket/743 (AbstractTripleStore.destroy() does not filter for correct prefix) +- http://trac.bigdata.com/ticket/746 (Assertion error) +- http://trac.bigdata.com/ticket/747 (BOUND bug) +- http://trac.bigdata.com/ticket/748 (incorrect join with subselect renaming vars) +- http://trac.bigdata.com/ticket/754 (Failure to setup SERVICE hook and changeLog for Unisolated and Read/Write connections) +- http://trac.bigdata.com/ticket/755 (Concurrent QuorumActors can interfere leading to failure to progress) +- http://trac.bigdata.com/ticket/756 (order by and group_concat) +- http://trac.bigdata.com/ticket/760 (Code review on 2-phase commit protocol) +- http://trac.bigdata.com/ticket/764 (RESYNC failure (HA)) +- http://trac.bigdata.com/ticket/770 (alpp ordering) +- http://trac.bigdata.com/ticket/772 (Query timeout only checked at operator start/stop.) +- http://trac.bigdata.com/ticket/776 (Closed as duplicate of #490) +- http://trac.bigdata.com/ticket/778 (HA Leader fail results in transient problem with allocations on other services) +- http://trac.bigdata.com/ticket/783 (Operator Alerts (HA)) 1.2.4: -- http://sourceforge.net/apps/trac/bigdata/ticket/777 (ConcurrentModificationException in ASTComplexOptionalOptimizer) +- http://trac.bigdata.com/ticket/777 (ConcurrentModificationException in ASTComplexOptionalOptimizer) 1.2.3: -- http://sourceforge.net/apps/trac/bigdata/ticket/168 (Maven Build) -- http://sourceforge.net/apps/trac/bigdata/ticket/196 (Journal leaks memory). -- http://sourceforge.net/apps/trac/bigdata/ticket/235 (Occasional deadlock in CI runs in com.bigdata.io.writecache.TestAll) -- http://sourceforge.net/apps/trac/bigdata/ticket/312 (CI (mock) quorums deadlock) -- http://sourceforge.net/apps/trac/bigdata/ticket/405 (Optimize hash join for subgroups with no incoming bound vars.) -- http://sourceforge.net/apps/trac/bigdata/ticket/412 (StaticAnalysis#getDefinitelyBound() ignores exogenous variables.) -- http://sourceforge.net/apps/trac/bigdata/ticket/485 (RDFS Plus Profile) -- http://sourceforge.net/apps/trac/bigdata/ticket/495 (SPARQL 1.1 Property Paths) -- http://sourceforge.net/apps/trac/bigdata/ticket/519 (Negative parser tests) -- http://sourceforge.net/apps/trac/bigdata/ticket/531 (SPARQL UPDATE for SOLUTION SETS) -- http://sourceforge.net/apps/trac/bigdata/ticket/535 (Optimize JOIN VARS for Sub-Selects) -- http://sourceforge.net/apps/trac/bigdata/ticket/555 (Support PSOutputStream/InputStream at IRawStore) -- http://sourceforge.net/apps/trac/bigdata/ticket/559 (Use RDFFormat.NQUADS as the format identifier for the NQuads parser) -- http://sourceforge.net/apps/trac/bigdata/ticket/570 (MemoryManager Journal does not implement all methods). -- http://sourceforge.net/apps/trac/bigdata/ticket/575 (NSS Admin API) -- http://sourceforge.net/apps/trac/bigdata/ticket/577 (DESCRIBE with OFFSET/LIMIT needs to use sub-select) -- http://sourceforge.net/apps/trac/bigdata/ticket/578 (Concise Bounded Description (CBD)) -- http://sourceforge.net/apps/trac/bigdata/ticket/579 (CONSTRUCT should use distinct SPO filter) -- http://sourceforge.net/apps/trac/bigdata/ticket/583 (VoID in ServiceDescription) -- http://sourceforge.net/apps/trac/bigdata/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) -- http://sourceforge.net/apps/trac/bigdata/ticket/590 (nxparser fails with uppercase language tag) -- http://sourceforge.net/apps/trac/bigdata/ticket/592 (Optimize RWStore allocator sizes) -- http://sourceforge.net/apps/trac/bigdata/ticket/593 (Ugrade to Sesame 2.6.10) -- http://sourceforge.net/apps/trac/bigdata/ticket/594 (WAR was deployed using TRIPLES rather than QUADS by default) -- http://sourceforge.net/apps/trac/bigdata/ticket/596 (Change web.xml parameter names to be consistent with Jini/River) -- http://sourceforge.net/apps/trac/bigdata/ticket/597 (SPARQL UPDATE LISTENER) -- http://sourceforge.net/apps/trac/bigdata/ticket/598 (B+Tree branching factor and HTree addressBits are confused in their NodeSerializer implementations) -- http://sourceforge.net/apps/trac/bigdata/ticket/599 (BlobIV for blank node : NotMaterializedException) -- http://sourceforge.net/apps/trac/bigdata/ticket/600 (BlobIV collision counter hits false limit.) -- http://sourceforge.net/apps/trac/bigdata/ticket/601 (Log uncaught exceptions) -- http://sourceforge.net/apps/trac/bigdata/ticket/602 (RWStore does not discard logged deletes on reset()) -- http://sourceforge.net/apps/trac/bigdata/ticket/607 (History service / index) -- http://sourceforge.net/apps/trac/bigdata/ticket/608 (LOG BlockingBuffer not progressing at INFO or lower level) -- http://sourceforge.net/apps/trac/bigdata/ticket/609 (bigdata-ganglia is required dependency for Journal) -- http://sourceforge.net/apps/trac/bigdata/ticket/611 (The code that processes SPARQL Update has a typo) -- http://sourceforge.net/apps/trac/bigdata/ticket/612 (Bigdata scale-up depends on zookeper) -- http://sourceforge.net/apps/trac/bigdata/ticket/613 (SPARQL UPDATE response inlines large DELETE or INSERT triple graphs) -- http://sourceforge.net/apps/trac/bigdata/ticket/614 (static join optimizer does not get ordering right when multiple tails share vars with ancestry) -- http://sourceforge.net/apps/trac/bigdata/ticket/615 (AST2BOpUtility wraps UNION with an unnecessary hash join) -- http://sourceforge.net/apps/trac/bigdata/ticket/616 (Row store read/update not isolated on Journal) -- http://sourceforge.net/apps/trac/bigdata/ticket/617 (Concurrent KB create fails with "No axioms defined?") -- http://sourceforge.net/apps/trac/bigdata/ticket/618 (DirectBufferPool.poolCapacity maximum of 2GB) -- http://sourceforge.net/apps/trac/bigdata/ticket/619 (RemoteRepository class should use application/x-www-form-urlencoded for large POST requests) -- http://sourceforge.net/apps/trac/bigdata/ticket/620 (UpdateServlet fails to parse MIMEType when doing conneg.) -- http://sourceforge.net/apps/trac/bigdata/ticket/626 (Expose performance counters for read-only indices) -- http://sourceforge.net/apps/trac/bigdata/ticket/627 (Environment variable override for NSS properties file) -- http://sourceforge.net/apps/trac/bigdata/ticket/628 (Create a bigdata-client jar for the NSS REST API) -- http://sourceforge.net/apps/trac/bigdata/ticket/631 (ClassCastException in SIDs mode query) -- http://sourceforge.net/apps/trac/bigdata/ticket/632 (NotMaterializedException when a SERVICE call needs variables that are provided as query input bindings) -- http://sourceforge.net/apps/trac/bigdata/ticket/633 (ClassCastException when binding non-uri values to a variable that occurs in predicate position) -- http://sourceforge.net/apps/trac/bigdata/ticket/638 (Change DEFAULT_MIN_RELEASE_AGE to 1ms) -- http://sourceforge.net/apps/trac/bigdata/ticket/640 (Conditionally rollback() BigdataSailConnection if dirty) -- http://sourceforge.net/apps/trac/bigdata/ticket/642 (Property paths do not work inside of exists/not exists filters) -- http://sourceforge.net/apps/trac/bigdata/ticket/643 (Add web.xml parameters to lock down public NSS end points) -- http://sourceforge.net/apps/trac/bigdata/ticket/644 (Bigdata2Sesame2BindingSetIterator can fail to notice asynchronous close()) -- http://sourceforge.net/apps/trac/bigdata/ticket/650 (Can not POST RDF to a graph using REST API) -- http://sourceforge.net/apps/trac/bigdata/ticket/654 (Rare AssertionError in WriteCache.clearAddrMap()) -- http://sourceforge.net/apps/trac/bigdata/ticket/655 (SPARQL REGEX operator does not perform case-folding correctly for Unicode data) -- http://sourceforge.net/apps/trac/bigdata/ticket/656 (InFactory bug when IN args consist of a single literal) -- http://sourceforge.net/apps/trac/bigdata/ticket/647 (SIDs mode creates unnecessary hash join for GRAPH group patterns) -- http://sourceforge.net/apps/trac/bigdata/ticket/667 (Provide NanoSparqlServer initialization hook) -- http://sourceforge.net/apps/trac/bigdata/ticket/669 (Doubly nested subqueries yield no results with LIMIT) -- http://sourceforge.net/apps/trac/bigdata/ticket/675 (Flush indices in parallel during checkpoint to reduce IO latency) -- http://sourceforge.net/apps/trac/bigdata/ticket/682 (AtomicRowFilter UnsupportedOperationException) +- http://trac.bigdata.com/ticket/168 (Maven Build) +- http://trac.bigdata.com/ticket/196 (Journal leaks memory). +- http://trac.bigdata.com/ticket/235 (Occasional deadlock in CI runs in com.bigdata.io.writecache.TestAll) +- http://trac.bigdata.com/ticket/312 (CI (mock) quorums deadlock) +- http://trac.bigdata.com/ticket/405 (Optimize hash join for subgroups with no incoming bound vars.) +- http://trac.bigdata.com/ticket/412 (StaticAnalysis#getDefinitelyBound() ignores exogenous variables.) +- http://trac.bigdata.com/ticket/485 (RDFS Plus Profile) +- http://trac.bigdata.com/ticket/495 (SPARQL 1.1 Property Paths) +- http://trac.bigdata.com/ticket/519 (Negative parser tests) +- http://trac.bigdata.com/ticket/531 (SPARQL UPDATE for SOLUTION SETS) +- http://trac.bigdata.com/ticket/535 (Optimize JOIN VARS for Sub-Selects) +- http://trac.bigdata.com/ticket/555 (Support PSOutputStream/InputStream at IRawStore) +- http://trac.bigdata.com/ticket/559 (Use RDFFormat.NQUADS as the format identifier for the NQuads parser) +- http://trac.bigdata.com/ticket/570 (MemoryManager Journal does not implement all methods). +- http://trac.bigdata.com/ticket/575 (NSS Admin API) +- http://trac.bigdata.com/ticket/577 (DESCRIBE with OFFSET/LIMIT needs to use sub-select) +- http://trac.bigdata.com/ticket/578 (Concise Bounded Description (CBD)) +- http://trac.bigdata.com/ticket/579 (CONSTRUCT should use distinct SPO filter) +- http://trac.bigdata.com/ticket/583 (VoID in ServiceDescription) +- http://trac.bigdata.com/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) +- http://trac.bigdata.com/ticket/590 (nxparser fails with uppercase language tag) +- http://trac.bigdata.com/ticket/592 (Optimize RWStore allocator sizes) +- http://trac.bigdata.com/ticket/593 (Ugrade to Sesame 2.6.10) +- http://trac.bigdata.com/ticket/594 (WAR was deployed using TRIPLES rather than QUADS by default) +- http://trac.bigdata.com/ticket/596 (Change web.xml parameter names to be consistent with Jini/River) +- http://trac.bigdata.com/ticket/597 (SPARQL UPDATE LISTENER) +- http://trac.bigdata.com/ticket/598 (B+Tree branching factor and HTree addressBits are confused in their NodeSerializer implementations) +- http://trac.bigdata.com/ticket/599 (BlobIV for blank node : NotMaterializedException) +- http://trac.bigdata.com/ticket/600 (BlobIV collision counter hits false limit.) +- http://trac.bigdata.com/ticket/601 (Log uncaught exceptions) +- http://trac.bigdata.com/ticket/602 (RWStore does not discard logged deletes on reset()) +- http://trac.bigdata.com/ticket/607 (History service / index) +- http://trac.bigdata.com/ticket/608 (LOG BlockingBuffer not progressing at INFO or lower level) +- http://trac.bigdata.com/ticket/609 (bigdata-ganglia is required dependency for Journal) +- http://trac.bigdata.com/ticket/611 (The code that processes SPARQL Update has a typo) +- http://trac.bigdata.com/ticket/612 (Bigdata scale-up depends on zookeper) +- http://trac.bigdata.com/ticket/613 (SPARQL UPDATE response inlines large DELETE or INSERT triple graphs) +- http://trac.bigdata.com/ticket/614 (static join optimizer does not get ordering right when multiple tails share vars with ancestry) +- http://trac.bigdata.com/ticket/615 (AST2BOpUtility wraps UNION with an unnecessary hash join) +- http://trac.bigdata.com/ticket/616 (Row store read/update not isolated on Journal) +- http://trac.bigdata.com/ticket/617 (Concurrent KB create fails with "No axioms defined?") +- http://trac.bigdata.com/ticket/618 (DirectBufferPool.poolCapacity maximum of 2GB) +- http://trac.bigdata.com/ticket/619 (RemoteRepository class should use application/x-www-form-urlencoded for large POST requests) +- http://trac.bigdata.com/ticket/620 (UpdateServlet fails to parse MIMEType when doing conneg.) +- http://trac.bigdata.com/ticket/626 (Expose performance counters for read-only indices) +- http://trac.bigdata.com/ticket/627 (Environment variable override for NSS properties file) +- http://trac.bigdata.com/ticket/628 (Create a bigdata-client jar for the NSS REST API) +- http://trac.bigdata.com/ticket/631 (ClassCastException in SIDs mode query) +- http://trac.bigdata.com/ticket/632 (NotMaterializedException when a SERVICE call needs variables that are provided as query input bindings) +- http://trac.bigdata.com/ticket/633 (ClassCastException when binding non-uri values to a variable that occurs in predicate position) +- http://trac.bigdata.com/ticket/638 (Change DEFAULT_MIN_RELEASE_AGE to 1ms) +- http://trac.bigdata.com/ticket/640 (Conditionally rollback() BigdataSailConnection if dirty) +- http://trac.bigdata.com/ticket/642 (Property paths do not work inside of exists/not exists filters) +- http://trac.bigdata.com/ticket/643 (Add web.xml parameters to lock down public NSS end points) +- http://trac.bigdata.com/ticket/644 (Bigdata2Sesame2BindingSetIterator can fail to notice asynchronous close()) +- http://trac.bigdata.com/ticket/650 (Can not POST RDF to a graph using REST API) +- http://trac.bigdata.com/ticket/654 (Rare AssertionError in WriteCache.clearAddrMap()) +- http://trac.bigdata.com/ticket/655 (SPARQL REGEX operator does not perform case-folding correctly for Unicode data) +- http://trac.bigdata.com/ticket/656 (InFactory bug when IN args consist of a single literal) +- http://trac.bigdata.com/ticket/647 (SIDs mode creates unnecessary hash join for GRAPH group patterns) +- http://trac.bigdata.com/ticket/667 (Provide NanoSparqlServer initialization hook) +- http://trac.bigdata.com/ticket/669 (Doubly nested subqueries yield no results with LIMIT) +- http://trac.bigdata.com/ticket/675 (Flush indices in parallel during checkpoint to reduce IO latency) +- http://trac.bigdata.com/ticket/682 (AtomicRowFilter UnsupportedOperationException) 1.2.2: -- http://sourceforge.net/apps/trac/bigdata/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) -- http://sourceforge.net/apps/trac/bigdata/ticket/602 (RWStore does not discard logged deletes on reset()) -- http://sourceforge.net/apps/trac/bigdata/ticket/603 (Prepare critical maintenance release as branch of 1.2.1) +- http://trac.bigdata.com/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) +- http://trac.bigdata.com/ticket/602 (RWStore does not discard logged deletes on reset()) +- http://trac.bigdata.com/ticket/603 (Prepare critical maintenance release as branch of 1.2.1) 1.2.1: -- http://sourceforge.net/apps/trac/bigdata/ticket/533 (Review materialization for inline IVs) -- http://sourceforge.net/apps/trac/bigdata/ticket/539 (NotMaterializedException with REGEX and Vocab) -- http://sourceforge.net/apps/trac/bigdata/ticket/540 (SPARQL UPDATE using NSS via index.html) -- http://sourceforge.net/apps/trac/bigdata/ticket/541 (MemoryManaged backed Journal mode) -- http://sourceforge.net/apps/trac/bigdata/ticket/546 (Index cache for Journal) -- http://sourceforge.net/apps/trac/bigdata/ticket/549 (BTree can not be cast to Name2Addr (MemStore recycler)) -- http://sourceforge.net/apps/trac/bigdata/ticket/550 (NPE in Leaf.getKey() : root cause was user error) -- http://sourceforge.net/apps/trac/bigdata/ticket/558 (SPARQL INSERT not working in same request after INSERT DATA) -- http://sourceforge.net/apps/trac/bigdata/ticket/562 (Sub-select in INSERT cause NPE in UpdateExprBuilder) -- http://sourceforge.net/apps/trac/bigdata/ticket/563 (DISTINCT ORDER BY) -- http://sourceforge.net/apps/trac/bigdata/ticket/567 (Failure to set cached value on IV results in incorrect behavior for complex UPDATE operation) -- http://sourceforge.net/apps/trac/bigdata/ticket/568 (DELETE WHERE fails with Java AssertionError) -- http://sourceforge.net/apps/trac/bigdata/ticket/569 (LOAD-CREATE-LOAD using virgin journal fails with "Graph exists" exception) -- http://sourceforge.net/apps/trac/bigdata/ticket/571 (DELETE/INSERT WHERE handling of blank nodes) -- http://sourceforge.net/apps/trac/bigdata/ticket/573 (NullPointerException when attempting to INSERT DATA containing a blank node) +- http://trac.bigdata.com/ticket/533 (Review materialization for inline IVs) +- http://trac.bigdata.com/ticket/539 (NotMaterializedException with REGEX and Vocab) +- http://trac.bigdata.com/ticket/540 (SPARQL UPDATE using NSS via index.html) +- http://trac.bigdata.com/ticket/541 (MemoryManaged backed Journal mode) +- http://trac.bigdata.com/ticket/546 (Index cache for Journal) +- http://trac.bigdata.com/ticket/549 (BTree can not be cast to Name2Addr (MemStore recycler)) +- http://trac.bigdata.com/ticket/550 (NPE in Leaf.getKey() : root cause was user error) +- http://trac.bigdata.com/ticket/558 (SPARQL INSERT not working in same request after INSERT DATA) +- http://trac.bigdata.com/ticket/562 (Sub-select in INSERT cause NPE in UpdateExprBuilder) +- http://trac.bigdata.com/ticket/563 (DISTINCT ORDER BY) +- http://trac.bigdata.com/ticket/567 (Failure to set cached value on IV results in incorrect behavior for complex UPDATE operation) +- http://trac.bigdata.com/ticket/568 (DELETE WHERE fails with Java AssertionError) +- http://trac.bigdata.com/ticket/569 (LOAD-CREATE-LOAD using virgin journal fails with "Graph exists" exception) +- http://trac.bigdata.com/ticket/571 (DELETE/INSERT WHERE handling of blank nodes) +- http://trac.bigdata.com/ticket/573 (NullPointerException when attempting to INSERT DATA containing a blank node) 1.2.0: (*) -- http://sourceforge.net/apps/trac/bigdata/ticket/92 (Monitoring webapp) -- http://sourceforge.net/apps/trac/bigdata/ticket/267 (Support evaluation of 3rd party operators) -- http://sourceforge.net/apps/trac/bigdata/ticket/337 (Compact and efficient movement of binding sets between nodes.) -- http://sourceforge.net/apps/trac/bigdata/ticket/433 (Cluster leaks threads under read-only index operations: DGC thread leak) -- http://sourceforge.net/apps/trac/bigdata/ticket/437 (Thread-local cache combined with unbounded thread pools causes effective memory leak: termCache memory leak & thread-local buffers) -- http://sourceforge.net/apps/trac/bigdata/ticket/438 (KeyBeforePartitionException on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/439 (Class loader problem) -- http://sourceforge.net/apps/trac/bigdata/ticket/441 (Ganglia integration) -- http://sourceforge.net/apps/trac/bigdata/ticket/443 (Logger for RWStore transaction service and recycler) -- http://sourceforge.net/apps/trac/bigdata/ticket/444 (SPARQL query can fail to notice when IRunningQuery.isDone() on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/445 (RWStore does not track tx release correctly) -- http://sourceforge.net/apps/trac/bigdata/ticket/446 (HTTP Repostory broken with bigdata 1.1.0) -- http://sourceforge.net/apps/trac/bigdata/ticket/448 (SPARQL 1.1 UPDATE) -- http://sourceforge.net/apps/trac/bigdata/ticket/449 (SPARQL 1.1 Federation extension) -- http://sourceforge.net/apps/trac/bigdata/ticket/451 (Serialization error in SIDs mode on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/454 (Global Row Store Read on Cluster uses Tx) -- http://sourceforge.net/apps/trac/bigdata/ticket/456 (IExtension implementations do point lookups on lexicon) -- http://sourceforge.net/apps/trac/bigdata/ticket/457 ("No such index" on cluster under concurrent query workload) -- http://sourceforge.net/apps/trac/bigdata/ticket/458 (Java level deadlock in DS) -- http://sourceforge.net/apps/trac/bigdata/ticket/460 (Uncaught interrupt resolving RDF terms) -- http://sourceforge.net/apps/trac/bigdata/ticket/461 (KeyAfterPartitionException / KeyBeforePartitionException on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/463 (NoSuchVocabularyItem with LUBMVocabulary for DerivedNumericsExtension) -- http://sourceforge.net/apps/trac/bigdata/ticket/464 (Query statistics do not update correctly on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/465 (Too many GRS reads on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/469 (Sail does not flush assertion buffers before query) -- http://sourceforge.net/apps/trac/bigdata/ticket/472 (acceptTaskService pool size on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/475 (Optimize serialization for query messages on cluster) -- http://sourceforge.net/apps/trac/bigdata/ticket/476 (Test suite for writeCheckpoint() and recycling for BTree/HTree) -- http://sourceforge.net/apps/trac/bigdata/ticket/478 (Cluster does not map input solution(s) across shards) -- http://sourceforge.net/apps/trac/bigdata/ticket/480 (Error releasing deferred frees using 1.0.6 against a 1.0.4 journal) -- http://sourceforge.net/apps/trac/bigdata/ticket/481 (PhysicalAddressResolutionException against 1.0.6) -- http://sourceforge.net/apps/trac/bigdata/ticket/482 (RWStore reset() should be thread-safe for concurrent readers) -- http://sourceforge.net/apps/trac/bigdata/ticket/484 (Java API for NanoSparqlServer REST API) -- http://sourceforge.net/apps/trac/bigdata/ticket/491 (AbstractTripleStore.destroy() does not clear the locator cache) -- http://sourceforge.net/apps/trac/bigdata/ticket/492 (Empty chunk in ThickChunkMessage (cluster)) -- http://sourceforge.net/apps/trac/bigdata/ticket/493 (Virtual Graphs) -- http://sourceforge.net/apps/trac/bigdata/ticket/496 (Sesame 2.6.3) -- http://sourceforge.net/apps/trac/bigdata/ticket/497 (Implement STRBEFORE, STRAFTER, and REPLACE) -- http://sourceforge.net/apps/trac/bigdata/ticket/498 (Bring bigdata RDF/XML parser up to openrdf 2.6.3.) -- http://sourceforge.net/apps/trac/bigdata/ticket/500 (SPARQL 1.1 Service Description) +- http://trac.bigdata.com/ticket/92 (Monitoring webapp) +- http://trac.bigdata.com/ticket/267 (Support evaluation of 3rd party operators) +- http://trac.bigdata.com/ticket/337 (Compact and efficient movement of binding sets between nodes.) +- http://trac.bigdata.com/ticket/433 (Cluster leaks threads under read-only index operations: DGC thread leak) +- http://trac.bigdata.com/ticket/437 (Thread-local cache combined with unbounded thread pools causes effective memory leak: termCache memory leak & thread-local buffers) +- http://trac.bigdata.com/ticket/438 (KeyBeforePartitionException on cluster) +- http://trac.bigdata.com/ticket/439 (Class loader problem) +- http://trac.bigdata.com/ticket/441 (Ganglia integration) +- http://trac.bigdata.com/ticket/443 (Logger for RWStore transaction service and recycler) +- http://trac.bigdata.com/ticket/444 (SPARQL query can fail to notice when IRunningQuery.isDone() on cluster) +- http://trac.bigdata.com/ticket/445 (RWStore does not track tx release correctly) +- http://trac.bigdata.com/ticket/446 (HTTP Repostory broken with bigdata 1.1.0) +- http://trac.bigdata.com/ticket/448 (SPARQL 1.1 UPDATE) +- http://trac.bigdata.com/ticket/449 (SPARQL 1.1 Federation extension) +- http://trac.bigdata.com/ticket/451 (Serialization error in SIDs mode on cluster) +- http://trac.bigdata.com/ticket/454 (Global Row Store Read on Cluster uses Tx) +- http://trac.bigdata.com/ticket/456 (IExtension implementations do point lookups on lexicon) +- http://trac.bigdata.com/ticket/457 ("No such index" on cluster under concurrent query workload) +- http://trac.bigdata.com/ticket/458 (Java level deadlock in DS) +- http://trac.bigdata.com/ticket/460 (Uncaught interrupt resolving RDF terms) +- http://trac.bigdata.com/ticket/461 (KeyAfterPartitionException / KeyBeforePartitionException on cluster) +- http://trac.bigdata.com/ticket/463 (NoSuchVocabularyItem with LUBMVocabulary for DerivedNumericsExtension) +- http://trac.bigdata.com/ticket/464 (Query statistics do not update correctly on cluster) +- http://trac.bigdata.com/ticket/465 (Too many GRS reads on cluster) +- http://trac.bigdata.com/ticket/469 (Sail does not flush assertion buffers before query) +- http://trac.bigdata.com/ticket/472 (acceptTaskService pool size on cluster) +- http://trac.bigdata.com/ticket/475 (Optimize serialization for query messages on cluster) +- http://trac.bigdata.com/ticket/476 (Test suite for writeCheckpoint() and recycling for BTree/HTree) +- http://trac.bigdata.com/ticket/478 (Cluster does not map input solution(s) across shards) +- http://trac.bigdata.com/ticket/480 (Error releasing deferred frees using 1.0.6 against a 1.0.4 journal) +- http://trac.bigdata.com/ticket/481 (PhysicalAddressResolutionException against 1.0.6) +- http://trac.bigdata.com/ticket/482 (RWStore reset() should be thread-safe for concurrent readers) +- http://trac.bigdata.com/ticket/484 (Java API for NanoSparqlServer REST API) +- http://trac.bigdata.com/ticket/491 (AbstractTripleStore.destroy() does not clear the locator cache) +- http://trac.bigdata.com/ticket/492 (Empty chunk in ThickChunkMessage (cluster)) +- http://trac.bigdata.com/ticket/493 (Virtual Graphs) +- http://trac.bigdata.com/ticket/496 (Sesame 2.6.3) +- http://trac.bigdata.com/ticket/497 (Implement STRBEFORE, STRAFTER, and REPLACE) +- http://trac.bigdata.com/ticket/498 (Bring bigdata RDF/XML parser up to openrdf 2.6.3.) +- http://trac.bigdata.com/ticket/500 (SPARQL 1.1 Service Description) - http://www.openrdf.org/issues/browse/SES-884 (Aggregation with an solution set as input should produce an empty solution as output) - http://www.openrdf.org/issues/browse/SES-862 (Incorrect error handling for SPARQL aggregation; fix in 2.6.1) - http://www.openrdf.org/issues/browse/SES-873 (Order the same Blank Nodes together in ORDER BY) -- http://sourceforge.net/apps/trac/bigdata/ticket/501 (SPARQL 1.1 BINDINGS are ignored) -- http://sourceforge.net/apps/trac/bigdata/ticket/503 (Bigdata2Sesame2BindingSetIterator throws QueryEvaluationException were it should throw NoSuchElementException) -- http://sourceforge.net/apps/trac/bigdata/ticket/504 (UNION with Empty Group Pattern) -- http://sourceforge.net/apps/trac/bigdata/ticket/505 (Exception when using SPARQL sort & statement identifiers) -- http://sourceforge.net/apps/trac/bigdata/ticket/506 (Load, closure and query performance in 1.1.x versus 1.0.x) -- http://sourceforge.net/apps/trac/bigdata/ticket/508 (LIMIT causes hash join utility to log errors) -- http://sourceforge.net/apps/trac/bigdata/ticket/513 (Expose the LexiconConfiguration to Function BOPs) -- http://sourceforge.net/apps/trac/bigdata/ticket/515 (Query with two "FILTER NOT EXISTS" expressions returns no results) -- http://sourceforge.net/apps/trac/bigdata/ticket/516 (REGEXBOp should cache the Pattern when it is a constant) -- http://sourceforge.net/apps/trac/bigdata/ticket/517 (Java 7 Compiler Compatibility) -- http://sourceforge.net/apps/trac/bigdata/ticket/518 (Review function bop subclass hierarchy, optimize datatype bop, etc.) -- http://sourceforge.net/apps/trac/bigdata/ticket/520 (CONSTRUCT WHERE shortcut) -- http://sourceforge.net/apps/trac/bigdata/ticket/521 (Incremental materialization of Tuple and Graph query results) -- http://sourceforge.net/apps/trac/bigdata/ticket/525 (Modify the IChangeLog interface to support multiple agents) -- http://sourceforge.net/apps/trac/bigdata/ticket/527 (Expose timestamp of LexiconRelation to function bops) -- http://sourceforge.net/apps/trac/bigdata/ticket/532 (ClassCastException during hash join (can not be cast to TermId)) -- http://sourceforge.net/apps/trac/bigdata/ticket/533 (Review materialization for inline IVs) -- http://sourceforge.net/apps/trac/bigdata/ticket/534 (BSBM BI Q5 error using MERGE JOIN) +- http://trac.bigdata.com/ticket/501 (SPARQL 1.1 BINDINGS are ignored) +- http://trac.bigdata.com/ticket/503 (Bigdata2Sesame2BindingSetIterator throws QueryEvaluationException were it should throw NoSuchElementException) +- http://trac.bigdata.com/ticket/504 (UNION with Empty Group Pattern) +- http://trac.bigdata.com/ticket/505 (Exception when using SPARQL sort & statement identifiers) +- http://trac.bigdata.com/ticket/506 (Load, closure and query performance in 1.1.x versus 1.0.x) +- http://trac.bigdata.com/ticket/508 (LIMIT causes hash join utility to log errors) +- http://trac.bigdata.com/ticket/513 (Expose the LexiconConfiguration to Function BOPs) +- http://trac.bigdata.com/ticket/515 (Query with two "FILTER NOT EXISTS" expressions returns no results) +- http://trac.bigdata.com/ticket/516 (REGEXBOp should cache the Pattern when it is a constant) +- http://trac.bigdata.com/ticket/517 (Java 7 Compiler Compatibility) +- http://trac.bigdata.com/ticket/518 (Review function bop subclass hierarchy, optimize datatype bop, etc.) +- http://trac.bigdata.com/ticket/520 (CONSTRUCT WHERE shortcut) +- http://trac.bigdata.com/ticket/521 (Incremental materialization of Tuple and Graph query results) +- http://trac.bigdata.com/ticket/525 (Modify the IChangeLog interface to support multiple agents) +- http://trac.bigdata.com/ticket/527 (Expose timestamp of LexiconRelation to function bops) +- http://trac.bigdata.com/ticket/532 (ClassCastException during hash join (can not be cast to TermId)) +- http://trac.bigdata.com/ticket/533 (Review materialization for inline IVs) +- http://trac.bigdata.com/ticket/534 (BSBM BI Q5 error using MERGE JOIN) 1.1.0 (*) - - http://sourceforge.net/apps/trac/bigdata/ticket/23 (Lexicon joins) - - http://sourceforge.net/apps/trac/bigdata/ticket/109 (Store large literals as "blobs") - - http://sourceforge.net/apps/trac/bigdata/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) - - http://sourceforge.net/apps/trac/bigdata/ticket/203 (Implement an persistence capable hash table to support analytic query) - - http://sourceforge.net/apps/trac/bigdata/ticket/209 (AccessPath should visit binding sets rather than elements for high level query.) - - http://sourceforge.net/apps/trac/bigdata/ticket/227 (SliceOp appears to be necessary when operator plan should suffice without) - - http://sourceforge.net/apps/trac/bigdata/ticket/232 (Bottom-up evaluation semantics). - - http://sourceforge.net/apps/trac/bigdata/ticket/246 (Derived xsd numeric data types must be inlined as extension types.) - - http://sourceforge.net/apps/trac/bigdata/ticket/254 (Revisit pruning of intermediate variable bindings during query execution) - - http://sourceforge.net/apps/trac/bigdata/ticket/261 (Lift conditions out of subqueries.) - - http://sourceforge.net/apps/trac/bigdata/ticket/300 (Native ORDER BY) - - http://sourceforge.net/apps/trac/bigdata/ticket/324 (Inline predeclared URIs and namespaces in 2-3 bytes) - - http://sourceforge.net/apps/trac/bigdata/ticket/330 (NanoSparqlServer does not locate "html" resources when run from jar) - - http://sourceforge.net/apps/trac/bigdata/ticket/334 (Support inlining of unicode data in the statement indices.) - - http://sourceforge.net/apps/trac/bigdata/ticket/364 (Scalable default graph evaluation) - - http://sourceforge.net/apps/trac/bigdata/ticket/368 (Prune variable bindings during query evaluation) - - http://sourceforge.net/apps/trac/bigdata/ticket/370 (Direct translation of openrdf AST to bigdata AST) - - http://sourceforge.net/apps/trac/bigdata/ticket/373 (Fix StrBOp and other IValueExpressions) - - http://sourceforge.net/apps/trac/bigdata/ticket/377 (Optimize OPTIONALs with multiple statement patterns.) - - http://sourceforge.net/apps/trac/bigdata/ticket/380 (Native SPARQL evaluation on cluster) - - http://sourceforge.net/apps/trac/bigdata/ticket/387 (Cluster does not compute closure) - - http://sourceforge.net/apps/trac/bigdata/ticket/395 (HTree hash join performance) - - http://sourceforge.net/apps/trac/bigdata/ticket/401 (inline xsd:unsigned datatypes) - - http://sourceforge.net/apps/trac/bigdata/ticket/408 (xsd:string cast fails for non-numeric data) - - http://sourceforge.net/apps/trac/bigdata/ticket/421 (New query hints model.) - - http://sourceforge.net/apps/trac/bigdata/ticket/431 (Use of read-only tx per query defeats cache on cluster) + - http://trac.bigdata.com/ticket/23 (Lexicon joins) + - http://trac.bigdata.com/ticket/109 (Store large literals as "blobs") + - http://trac.bigdata.com/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) + - http://trac.bigdata.com/ticket/203 (Implement an persistence capable hash table to support analytic query) + - http://trac.bigdata.com/ticket/209 (AccessPath should visit binding sets rather than elements for high level query.) + - http://trac.bigdata.com/ticket/227 (SliceOp appears to be necessary when operator plan should suffice without) + - http://trac.bigdata.com/ticket/232 (Bottom-up evaluation semantics). + - http://trac.bigdata.com/ticket/246 (Derived xsd numeric data types must be inlined as extension types.) + - http://trac.bigdata.com/ticket/254 (Revisit pruning of intermediate variable bindings during query execution) + - http://trac.bigdata.com/ticket/261 (Lift conditions out of subqueries.) + - http://trac.bigdata.com/ticket/300 (Native ORDER BY) + - http://trac.bigdata.com/ticket/324 (Inline predeclared URIs and namespaces in 2-3 bytes) + - http://trac.bigdata.com/ticket/330 (NanoSparqlServer does not locate "html" resources when run from jar) + - http://trac.bigdata.com/ticket/334 (Support inlining of unicode data in the statement indices.) + - http://trac.bigdata.com/ticket/364 (Scalable default graph evaluation) + - http://trac.bigdata.com/ticket/368 (Prune variable bindings during query evaluation) + - http://trac.bigdata.com/ticket/370 (Direct translation of openrdf AST to bigdata AST) + - http://trac.bigdata.com/ticket/373 (Fix StrBOp and other IValueExpressions) + - http://trac.bigdata.com/ticket/377 (Optimize OPTIONALs with multiple statement patterns.) + - http://trac.bigdata.com/ticket/380 (Native SPARQL evaluation on cluster) + - http://trac.bigdata.com/ticket/387 (Cluster does not compute closure) + - http://trac.bigdata.com/ticket/395 (HTree hash join performance) + - http://trac.bigdata.com/ticket/401 (inline xsd:unsigned datatypes) + - http://trac.bigdata.com/ticket/408 (xsd:string cast fails for non-numeric data) + - http://trac.bigdata.com/ticket/421 (New query hints model.) + - http://trac.bigdata.com/ticket/431 (Use of read-only tx per query defeats cache on cluster) 1.0.3 - - http://sourceforge.net/apps/trac/bigdata/ticket/217 (BTreeCounters does not track bytes released) - - http://sourceforge.net/apps/trac/bigdata/ticket/269 (Refactor performance counters using accessor interface) - - http://sourceforge.net/apps/trac/bigdata/ticket/329 (B+Tree should delete bloom filter when it is disabled.) - - http://sourceforge.net/apps/trac/bigdata/ticket/372 (RWStore does not prune the CommitRecordIndex) - - http://sourceforge.net/apps/trac/bigdata/ticket/375 (Persistent memory leaks (RWStore/DISK)) - - http://sourceforge.net/apps/trac/bigdata/ticket/385 (FastRDFValueCoder2: ArrayIndexOutOfBoundsException) - - http://sourceforge.net/apps/trac/bigdata/ticket/391 (Release age advanced on WORM mode journal) - - http://sourceforge.net/apps/trac/bigdata/ticket/392 (Add a DELETE by access path method to the NanoSparqlServer) - - http://sourceforge.net/apps/trac/bigdata/ticket/393 (Add "context-uri" request parameter to specify the default context for INSERT in the REST API) - - http://sourceforge.net/apps/trac/bigdata/ticket/394 (log4j configuration error message in WAR deployment) - - http://sourceforge.net/apps/trac/bigdata/ticket/399 (Add a fast range count method to the REST API) - - http://sourceforge.net/apps/trac/bigdata/ticket/422 (Support temp triple store wrapped by a BigdataSail) - - http://sourceforge.net/apps/trac/bigdata/ticket/424 (NQuads support for NanoSparqlServer) - - http://sourceforge.net/apps/trac/bigdata/ticket/425 (Bug fix to DEFAULT_RDF_FORMAT for bulk data loader in scale-out) - - http://sourceforge.net/apps/trac/bigdata/ticket/426 (Support either lockfile (procmail) and dotlockfile (liblockfile1) in scale-out) - - http://sourceforge.net/apps/trac/bigdata/ticket/427 (BigdataSail#getReadOnlyConnection() race condition with concurrent commit) - - http://sourceforge.net/apps/trac/bigdata/ticket/435 (Address is 0L) - - http://sourceforge.net/apps/trac/bigdata/ticket/436 (TestMROWTransactions failure in CI) + - http://trac.bigdata.com/ticket/217 (BTreeCounters does not track bytes released) + - http://trac.bigdata.com/ticket/269 (Refactor performance counters using accessor interface) + - http://trac.bigdata.com/ticket/329 (B+Tree should delete bloom filter when it is disabled.) + - http://trac.bigdata.com/ticket/372 (RWStore does not prune the CommitRecordIndex) + - http://trac.bigdata.com/ticket/375 (Persistent memory leaks (RWStore/DISK)) + - http://trac.bigdata.com/ticket/385 (FastRDFValueCoder2: ArrayIndexOutOfBoundsException) + - http://trac.bigdata.com/ticket/391 (Release age advanced on WORM mode journal) + - http://trac.bigdata.com/ticket/392 (Add a DELETE by access path method to the NanoSparqlServer) + - http://trac.bigdata.com/ticket/393 (Add "context-uri" request parameter to specify the default context for INSERT in the REST API) + - http://trac.bigdata.com/ticket/394 (log4j configuration error message in WAR deployment) + - http://trac.bigdata.com/ticket/399 (Add a fast range count method to the REST API) + - http://trac.bigdata.com/ticket/422 (Support temp triple store wrapped by a BigdataSail) + - http://trac.bigdata.com/ticket/424 (NQuads support for NanoSparqlServer) + - http://trac.bigdata.com/ticket/425 (Bug fix to DEFAULT_RDF_FORMAT for bulk data loader in scale-out) + - http://trac.bigdata.com/ticket/426 (Support either lockfile (procmail) and dotlockfile (liblockfile1) in scale-out) + - http://trac.bigdata.com/ticket/427 (BigdataSail#getReadOnlyConnection() race condition with concurrent commit) + - http://trac.bigdata.com/ticket/435 (Address is 0L) + - http://trac.bigdata.com/ticket/436 (TestMROWTransactions failure in CI) 1.0.2 - - http://sourceforge.net/apps/trac/bigdata/ticket/32 (Query time expansion of (foo rdf:type rdfs:Resource) drags in SPORelation for scale-out.) - - http://sourceforge.net/apps/trac/bigdata/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) - - http://sourceforge.net/apps/trac/bigdata/ticket/356 (Query not terminated by error.) - - http://sourceforge.net/apps/trac/bigdata/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) - - http://sourceforge.net/apps/trac/bigdata/ticket/361 (IRunningQuery not closed promptly.) - - http://sourceforge.net/apps/trac/bigdata/ticket/371 (DataLoader fails to load resources available from the classpath.) - - http://sourceforge.net/apps/trac/bigdata/ticket/376 (Support for the streaming of bigdata IBindingSets into a sparql query.) - - http://sourceforge.net/apps/trac/bigdata/ticket/378 (ClosedByInterruptException during heavy query mix.) - - http://sourceforge.net/apps/trac/bigdata/ticket/379 (NotSerializableException for SPOAccessPath.) - - http://sourceforge.net/apps/trac/bigdata/ticket/382 (Change dependencies to Apache River 2.2.0) + - http://trac.bigdata.com/ticket/32 (Query time expansion of (foo rdf:type rdfs:Resource) drags in SPORelation for scale-out.) + - http://trac.bigdata.com/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) + - http://trac.bigdata.com/ticket/356 (Query not terminated by error.) + - http://trac.bigdata.com/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) + - http://trac.bigdata.com/ticket/361 (IRunningQuery not closed promptly.) + - http://trac.bigdata.com/ticket/371 (DataLoader fails to load resources available from the classpath.) + - http://trac.bigdata.com/ticket/376 (Support for the streaming of bigdata IBindingSets into a sparql query.) + - http://trac.bigdata.com/ticket/378 (ClosedByInterruptException during heavy query mix.) + - http://trac.bigdata.com/ticket/379 (NotSerializableException for SPOAccessPath.) + - http://trac.bigdata.com/ticket/382 (Change dependencies to Apache River 2.2.0) 1.0.1 (*) - - http://sourceforge.net/apps/trac/bigdata/ticket/107 (Unicode clean schema names in the sparse row store). - - http://sourceforge.net/apps/trac/bigdata/ticket/124 (TermIdEncoder should use more bits for scale-out). - - http://sourceforge.net/apps/trac/bigdata/ticket/225 (OSX requires specialized performance counter collection classes). - - http://sourceforge.net/apps/trac/bigdata/ticket/348 (BigdataValueFactory.asValue() must return new instance when DummyIV is used). - - http://sourceforge.net/apps/trac/bigdata/ticket/349 (TermIdEncoder limits Journal to 2B distinct RDF Values per triple/quad store instance). - - http://sourceforge.net/apps/trac/bigdata/ticket/351 (SPO not Serializable exception in SIDS mode (scale-out)). - - http://sourceforge.net/apps/trac/bigdata/ticket/352 (ClassCastException when querying with binding-values that are not known to the database). - - http://sourceforge.net/apps/trac/bigdata/ticket/353 (UnsupportedOperatorException for some SPARQL queries). - - http://sourceforge.net/apps/trac/bigdata/ticket/355 (Query failure when comparing with non materialized value). - - http://sourceforge.net/apps/trac/bigdata/ticket/357 (RWStore reports "FixedAllocator returning null address, with freeBits".) - - http://sourceforge.net/apps/trac/bigdata/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) - - http://sourceforge.net/apps/trac/bigdata/ticket/362 (log4j - slf4j bridge.) + - http://trac.bigdata.com/ticket/107 (Unicode clean schema names in the sparse row store). + - http://trac.bigdata.com/ticket/124 (TermIdEncoder should use more bits for scale-out). + - http://trac.bigdata.com/ticket/225 (OSX requires specialized performance counter collection classes). + - http://trac.bigdata.com/ticket/348 (BigdataValueFactory.asValue() must return new instance when DummyIV is used). + - http://trac.bigdata.com/ticket/349 (TermIdEncoder limits Journal to 2B distinct RDF Values per triple/quad store instance). + - http://trac.bigdata.com/ticket/351 (SPO not Serializable exception in SIDS mode (scale-out)). + - http://trac.bigdata.com/ticket/352 (ClassCastException when querying with binding-values that are not known to the database). + - http://trac.bigdata.com/ticket/353 (UnsupportedOperatorException for some SPARQL queries). + - http://trac.bigdata.com/ticket/355 (Query failure when comparing with non materialized value). + - http://trac.bigdata.com/ticket/357 (RWStore reports "FixedAllocator returning null address, with freeBits".) + - http://trac.bigdata.com/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) + - http://trac.bigdata.com/ticket/362 (log4j - slf4j bridge.) For more information about bigdata(R), please see the following links: -[1] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=Main_Page -[2] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=GettingStarted -[3] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=Roadmap +[1] http://wiki.bigdata.com/wiki/index.php/Main_Page +[2] http://wiki.bigdata.com/wiki/index.php/GettingStarted +[3] http://wiki.bigdata.com/wiki/index.php/Roadmap [4] http://www.bigdata.com/bigdata/docs/api/ [5] http://sourceforge.net/projects/bigdata/ [6] http://www.bigdata.com/blog [7] http://www.systap.com/bigdata.htm [8] http://sourceforge.net/projects/bigdata/files/bigdata/ -[9] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=DataMigration -[10] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=HAJournalServer +[9] http://wiki.bigdata.com/wiki/index.php/DataMigration +[10] http://wiki.bigdata.com/wiki/index.php/HAJournalServer [11] http://www.bigdata.com/whitepapers/reifSPARQL.pdf -[12] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=RDF_GAS_API +[12] http://wiki.bigdata.com/wiki/index.php/RDF_GAS_API About bigdata: This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-05 19:01:07
|
Revision: 8202 http://sourceforge.net/p/bigdata/code/8202 Author: thompsonbry Date: 2014-05-05 19:01:03 +0000 (Mon, 05 May 2014) Log Message: ----------- Commit of release notes template for 1.3.1. Added Paths: ----------- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt Added: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/releases/RELEASE_1_3_1.txt 2014-05-05 19:01:03 UTC (rev 8202) @@ -0,0 +1,390 @@ +This is a major release of bigdata(R). + +Bigdata is a horizontally-scaled, open-source architecture for indexed data with an emphasis on RDF capable of loading 1B triples in under one hour on a 15 node cluster. Bigdata operates in both a single machine mode (Journal), highly available replication cluster mode (HAJournalServer), and a horizontally sharded cluster mode (BigdataFederation). The Journal provides fast scalable ACID indexed storage for very large data sets, up to 50 billion triples / quads. The HAJournalServer adds replication, online backup, horizontal scaling of query, and high availability. The federation provides fast scalable shard-wise parallel indexed storage using dynamic sharding and shard-wise ACID updates and incremental cluster size growth. Both platforms support fully concurrent readers with snapshot isolation. + +Distributed processing offers greater throughput but does not reduce query or update latency. Choose the Journal when the anticipated scale and throughput requirements permit. Choose the HAJournalServer for high availability and linear scaling in query throughput. Choose the BigdataFederation when the administrative and machine overhead associated with operating a cluster is an acceptable tradeoff to have essentially unlimited data scaling and throughput. + +See [1,2,8] for instructions on installing bigdata(R), [4] for the javadoc, and [3,5,6] for news, questions, and the latest developments. For more information about SYSTAP, LLC and bigdata, see [7]. + +Starting with the 1.0.0 release, we offer a WAR artifact [8] for easy installation of the single machine RDF database. For custom development and cluster installations we recommend checking out the code from SVN using the tag for this release. The code will build automatically under eclipse. You can also build the code using the ant script. The cluster installer requires the use of the ant script. + +Starting with the 1.3.0 release, we offer a tarball artifact [10] for easy installation of the HA replication cluster. + +You can download the WAR (standalone) or HA artifacts from: + +http://sourceforge.net/projects/bigdata/ + +You can checkout this release from: + +https://bigdata.svn.sourceforge.net/svnroot/bigdata/tags/BIGDATA_RELEASE_1_3_1 + +New features: + +- High availability [10]. +- HA Load balancer. +- New workbench. +- Blueprints API. +- Reification Done Right (RDR) support [11]. +- RDF Graph Mining Service (GASService) [12]. +- Property Path performance enhancements. +- Plus numerous other bug fixes and performance enhancements. + +Feature summary: + +- Highly Available Replication Clusters (HAJournalServer [10]) +- Single machine data storage to ~50B triples/quads (RWStore); +- Clustered data storage is essentially unlimited (BigdataFederation); +- Simple embedded and/or webapp deployment (NanoSparqlServer); +- Triples, quads, or triples with provenance (SIDs); +- Fast RDFS+ inference and truth maintenance; +- Fast 100% native SPARQL 1.1 evaluation; +- Integrated "analytic" query package; +- %100 Java memory manager leverages the JVM native heap (no GC); + +Road map [3]: + +- Column-wise indexing; +- Runtime Query Optimizer for Analytic Query mode; +- Performance optimization for scale-out clusters; and +- Simplified deployment, configuration, and administration for scale-out clusters. + +Change log: + + Note: Versions with (*) MAY require data migration. For details, see [9]. + +1.3.1: + +TBD + +1.3.0: + +- http://sourceforge.net/apps/trac/bigdata/ticket/530 (Journal HA) +- http://sourceforge.net/apps/trac/bigdata/ticket/621 (Coalesce write cache records and install reads in cache) +- http://sourceforge.net/apps/trac/bigdata/ticket/623 (HA TXS) +- http://sourceforge.net/apps/trac/bigdata/ticket/639 (Remove triple-buffering in RWStore) +- http://sourceforge.net/apps/trac/bigdata/ticket/645 (HA backup) +- http://sourceforge.net/apps/trac/bigdata/ticket/646 (River not compatible with newer 1.6.0 and 1.7.0 JVMs) +- http://sourceforge.net/apps/trac/bigdata/ticket/648 (Add a custom function to use full text index for filtering.) +- http://sourceforge.net/apps/trac/bigdata/ticket/651 (RWS test failure) +- http://sourceforge.net/apps/trac/bigdata/ticket/652 (Compress write cache blocks for replication and in HALogs) +- http://sourceforge.net/apps/trac/bigdata/ticket/662 (Latency on followers during commit on leader) +- http://sourceforge.net/apps/trac/bigdata/ticket/663 (Issue with OPTIONAL blocks) +- http://sourceforge.net/apps/trac/bigdata/ticket/664 (RWStore needs post-commit protocol) +- http://sourceforge.net/apps/trac/bigdata/ticket/665 (HA3 LOAD non-responsive with node failure) +- http://sourceforge.net/apps/trac/bigdata/ticket/666 (Occasional CI deadlock in HALogWriter testConcurrentRWWriterReader) +- http://sourceforge.net/apps/trac/bigdata/ticket/670 (Accumulating HALog files cause latency for HA commit) +- http://sourceforge.net/apps/trac/bigdata/ticket/671 (Query on follower fails during UPDATE on leader) +- http://sourceforge.net/apps/trac/bigdata/ticket/673 (DGC in release time consensus protocol causes native thread leak in HAJournalServer at each commit) +- http://sourceforge.net/apps/trac/bigdata/ticket/674 (WCS write cache compaction causes errors in RWS postHACommit()) +- http://sourceforge.net/apps/trac/bigdata/ticket/676 (Bad patterns for timeout computations) +- http://sourceforge.net/apps/trac/bigdata/ticket/677 (HA deadlock under UPDATE + QUERY) +- http://sourceforge.net/apps/trac/bigdata/ticket/678 (DGC Thread and Open File Leaks: sendHALogForWriteSet()) +- http://sourceforge.net/apps/trac/bigdata/ticket/679 (HAJournalServer can not restart due to logically empty log file) +- http://sourceforge.net/apps/trac/bigdata/ticket/681 (HAJournalServer deadlock: pipelineRemove() and getLeaderId()) +- http://sourceforge.net/apps/trac/bigdata/ticket/684 (Optimization with skos altLabel) +- http://sourceforge.net/apps/trac/bigdata/ticket/686 (Consensus protocol does not detect clock skew correctly) +- http://sourceforge.net/apps/trac/bigdata/ticket/687 (HAJournalServer Cache not populated) +- http://sourceforge.net/apps/trac/bigdata/ticket/689 (Missing URL encoding in RemoteRepositoryManager) +- http://sourceforge.net/apps/trac/bigdata/ticket/690 (Error when using the alias "a" instead of rdf:type for a multipart insert) +- http://sourceforge.net/apps/trac/bigdata/ticket/691 (Failed to re-interrupt thread in HAJournalServer) +- http://sourceforge.net/apps/trac/bigdata/ticket/692 (Failed to re-interrupt thread) +- http://sourceforge.net/apps/trac/bigdata/ticket/693 (OneOrMorePath SPARQL property path expression ignored) +- http://sourceforge.net/apps/trac/bigdata/ticket/694 (Transparently cancel update/query in RemoteRepository) +- http://sourceforge.net/apps/trac/bigdata/ticket/695 (HAJournalServer reports "follower" but is in SeekConsensus and is not participating in commits.) +- http://sourceforge.net/apps/trac/bigdata/ticket/701 (Problems in BackgroundTupleResult) +- http://sourceforge.net/apps/trac/bigdata/ticket/702 (InvocationTargetException on /namespace call) +- http://sourceforge.net/apps/trac/bigdata/ticket/704 (ask does not return json) +- http://sourceforge.net/apps/trac/bigdata/ticket/705 (Race between QueryEngine.putIfAbsent() and shutdownNow()) +- http://sourceforge.net/apps/trac/bigdata/ticket/706 (MultiSourceSequentialCloseableIterator.nextSource() can throw NPE) +- http://sourceforge.net/apps/trac/bigdata/ticket/707 (BlockingBuffer.close() does not unblock threads) +- http://sourceforge.net/apps/trac/bigdata/ticket/708 (BIND heisenbug - race condition on select query with BIND) +- http://sourceforge.net/apps/trac/bigdata/ticket/711 (sparql protocol: mime type application/sparql-query) +- http://sourceforge.net/apps/trac/bigdata/ticket/712 (SELECT ?x { OPTIONAL { ?x eg:doesNotExist eg:doesNotExist } } incorrect) +- http://sourceforge.net/apps/trac/bigdata/ticket/715 (Interrupt of thread submitting a query for evaluation does not always terminate the AbstractRunningQuery) +- http://sourceforge.net/apps/trac/bigdata/ticket/716 (Verify that IRunningQuery instances (and nested queries) are correctly cancelled when interrupted) +- http://sourceforge.net/apps/trac/bigdata/ticket/718 (HAJournalServer needs to handle ZK client connection loss) +- http://sourceforge.net/apps/trac/bigdata/ticket/720 (HA3 simultaneous service start failure) +- http://sourceforge.net/apps/trac/bigdata/ticket/723 (HA asynchronous tasks must be canceled when invariants are changed) +- http://sourceforge.net/apps/trac/bigdata/ticket/725 (FILTER EXISTS in subselect) +- http://sourceforge.net/apps/trac/bigdata/ticket/726 (Logically empty HALog for committed transaction) +- http://sourceforge.net/apps/trac/bigdata/ticket/727 (DELETE/INSERT fails with OPTIONAL non-matching WHERE) +- http://sourceforge.net/apps/trac/bigdata/ticket/728 (Refactor to create HAClient) +- http://sourceforge.net/apps/trac/bigdata/ticket/729 (ant bundleJar not working) +- http://sourceforge.net/apps/trac/bigdata/ticket/731 (CBD and Update leads to 500 status code) +- http://sourceforge.net/apps/trac/bigdata/ticket/732 (describe statement limit does not work) +- http://sourceforge.net/apps/trac/bigdata/ticket/733 (Range optimizer not optimizing Slice service) +- http://sourceforge.net/apps/trac/bigdata/ticket/734 (two property paths interfere) +- http://sourceforge.net/apps/trac/bigdata/ticket/736 (MIN() malfunction) +- http://sourceforge.net/apps/trac/bigdata/ticket/737 (class cast exception) +- http://sourceforge.net/apps/trac/bigdata/ticket/739 (Inconsistent treatment of bind and optional property path) +- http://sourceforge.net/apps/trac/bigdata/ticket/741 (ctc-striterators should build as independent top-level project (Apache2)) +- http://sourceforge.net/apps/trac/bigdata/ticket/743 (AbstractTripleStore.destroy() does not filter for correct prefix) +- http://sourceforge.net/apps/trac/bigdata/ticket/746 (Assertion error) +- http://sourceforge.net/apps/trac/bigdata/ticket/747 (BOUND bug) +- http://sourceforge.net/apps/trac/bigdata/ticket/748 (incorrect join with subselect renaming vars) +- http://sourceforge.net/apps/trac/bigdata/ticket/754 (Failure to setup SERVICE hook and changeLog for Unisolated and Read/Write connections) +- http://sourceforge.net/apps/trac/bigdata/ticket/755 (Concurrent QuorumActors can interfere leading to failure to progress) +- http://sourceforge.net/apps/trac/bigdata/ticket/756 (order by and group_concat) +- http://sourceforge.net/apps/trac/bigdata/ticket/760 (Code review on 2-phase commit protocol) +- http://sourceforge.net/apps/trac/bigdata/ticket/764 (RESYNC failure (HA)) +- http://sourceforge.net/apps/trac/bigdata/ticket/770 (alpp ordering) +- http://sourceforge.net/apps/trac/bigdata/ticket/772 (Query timeout only checked at operator start/stop.) +- http://sourceforge.net/apps/trac/bigdata/ticket/776 (Closed as duplicate of #490) +- http://sourceforge.net/apps/trac/bigdata/ticket/778 (HA Leader fail results in transient problem with allocations on other services) +- http://sourceforge.net/apps/trac/bigdata/ticket/783 (Operator Alerts (HA)) + +1.2.4: + +- http://sourceforge.net/apps/trac/bigdata/ticket/777 (ConcurrentModificationException in ASTComplexOptionalOptimizer) + +1.2.3: + +- http://sourceforge.net/apps/trac/bigdata/ticket/168 (Maven Build) +- http://sourceforge.net/apps/trac/bigdata/ticket/196 (Journal leaks memory). +- http://sourceforge.net/apps/trac/bigdata/ticket/235 (Occasional deadlock in CI runs in com.bigdata.io.writecache.TestAll) +- http://sourceforge.net/apps/trac/bigdata/ticket/312 (CI (mock) quorums deadlock) +- http://sourceforge.net/apps/trac/bigdata/ticket/405 (Optimize hash join for subgroups with no incoming bound vars.) +- http://sourceforge.net/apps/trac/bigdata/ticket/412 (StaticAnalysis#getDefinitelyBound() ignores exogenous variables.) +- http://sourceforge.net/apps/trac/bigdata/ticket/485 (RDFS Plus Profile) +- http://sourceforge.net/apps/trac/bigdata/ticket/495 (SPARQL 1.1 Property Paths) +- http://sourceforge.net/apps/trac/bigdata/ticket/519 (Negative parser tests) +- http://sourceforge.net/apps/trac/bigdata/ticket/531 (SPARQL UPDATE for SOLUTION SETS) +- http://sourceforge.net/apps/trac/bigdata/ticket/535 (Optimize JOIN VARS for Sub-Selects) +- http://sourceforge.net/apps/trac/bigdata/ticket/555 (Support PSOutputStream/InputStream at IRawStore) +- http://sourceforge.net/apps/trac/bigdata/ticket/559 (Use RDFFormat.NQUADS as the format identifier for the NQuads parser) +- http://sourceforge.net/apps/trac/bigdata/ticket/570 (MemoryManager Journal does not implement all methods). +- http://sourceforge.net/apps/trac/bigdata/ticket/575 (NSS Admin API) +- http://sourceforge.net/apps/trac/bigdata/ticket/577 (DESCRIBE with OFFSET/LIMIT needs to use sub-select) +- http://sourceforge.net/apps/trac/bigdata/ticket/578 (Concise Bounded Description (CBD)) +- http://sourceforge.net/apps/trac/bigdata/ticket/579 (CONSTRUCT should use distinct SPO filter) +- http://sourceforge.net/apps/trac/bigdata/ticket/583 (VoID in ServiceDescription) +- http://sourceforge.net/apps/trac/bigdata/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) +- http://sourceforge.net/apps/trac/bigdata/ticket/590 (nxparser fails with uppercase language tag) +- http://sourceforge.net/apps/trac/bigdata/ticket/592 (Optimize RWStore allocator sizes) +- http://sourceforge.net/apps/trac/bigdata/ticket/593 (Ugrade to Sesame 2.6.10) +- http://sourceforge.net/apps/trac/bigdata/ticket/594 (WAR was deployed using TRIPLES rather than QUADS by default) +- http://sourceforge.net/apps/trac/bigdata/ticket/596 (Change web.xml parameter names to be consistent with Jini/River) +- http://sourceforge.net/apps/trac/bigdata/ticket/597 (SPARQL UPDATE LISTENER) +- http://sourceforge.net/apps/trac/bigdata/ticket/598 (B+Tree branching factor and HTree addressBits are confused in their NodeSerializer implementations) +- http://sourceforge.net/apps/trac/bigdata/ticket/599 (BlobIV for blank node : NotMaterializedException) +- http://sourceforge.net/apps/trac/bigdata/ticket/600 (BlobIV collision counter hits false limit.) +- http://sourceforge.net/apps/trac/bigdata/ticket/601 (Log uncaught exceptions) +- http://sourceforge.net/apps/trac/bigdata/ticket/602 (RWStore does not discard logged deletes on reset()) +- http://sourceforge.net/apps/trac/bigdata/ticket/607 (History service / index) +- http://sourceforge.net/apps/trac/bigdata/ticket/608 (LOG BlockingBuffer not progressing at INFO or lower level) +- http://sourceforge.net/apps/trac/bigdata/ticket/609 (bigdata-ganglia is required dependency for Journal) +- http://sourceforge.net/apps/trac/bigdata/ticket/611 (The code that processes SPARQL Update has a typo) +- http://sourceforge.net/apps/trac/bigdata/ticket/612 (Bigdata scale-up depends on zookeper) +- http://sourceforge.net/apps/trac/bigdata/ticket/613 (SPARQL UPDATE response inlines large DELETE or INSERT triple graphs) +- http://sourceforge.net/apps/trac/bigdata/ticket/614 (static join optimizer does not get ordering right when multiple tails share vars with ancestry) +- http://sourceforge.net/apps/trac/bigdata/ticket/615 (AST2BOpUtility wraps UNION with an unnecessary hash join) +- http://sourceforge.net/apps/trac/bigdata/ticket/616 (Row store read/update not isolated on Journal) +- http://sourceforge.net/apps/trac/bigdata/ticket/617 (Concurrent KB create fails with "No axioms defined?") +- http://sourceforge.net/apps/trac/bigdata/ticket/618 (DirectBufferPool.poolCapacity maximum of 2GB) +- http://sourceforge.net/apps/trac/bigdata/ticket/619 (RemoteRepository class should use application/x-www-form-urlencoded for large POST requests) +- http://sourceforge.net/apps/trac/bigdata/ticket/620 (UpdateServlet fails to parse MIMEType when doing conneg.) +- http://sourceforge.net/apps/trac/bigdata/ticket/626 (Expose performance counters for read-only indices) +- http://sourceforge.net/apps/trac/bigdata/ticket/627 (Environment variable override for NSS properties file) +- http://sourceforge.net/apps/trac/bigdata/ticket/628 (Create a bigdata-client jar for the NSS REST API) +- http://sourceforge.net/apps/trac/bigdata/ticket/631 (ClassCastException in SIDs mode query) +- http://sourceforge.net/apps/trac/bigdata/ticket/632 (NotMaterializedException when a SERVICE call needs variables that are provided as query input bindings) +- http://sourceforge.net/apps/trac/bigdata/ticket/633 (ClassCastException when binding non-uri values to a variable that occurs in predicate position) +- http://sourceforge.net/apps/trac/bigdata/ticket/638 (Change DEFAULT_MIN_RELEASE_AGE to 1ms) +- http://sourceforge.net/apps/trac/bigdata/ticket/640 (Conditionally rollback() BigdataSailConnection if dirty) +- http://sourceforge.net/apps/trac/bigdata/ticket/642 (Property paths do not work inside of exists/not exists filters) +- http://sourceforge.net/apps/trac/bigdata/ticket/643 (Add web.xml parameters to lock down public NSS end points) +- http://sourceforge.net/apps/trac/bigdata/ticket/644 (Bigdata2Sesame2BindingSetIterator can fail to notice asynchronous close()) +- http://sourceforge.net/apps/trac/bigdata/ticket/650 (Can not POST RDF to a graph using REST API) +- http://sourceforge.net/apps/trac/bigdata/ticket/654 (Rare AssertionError in WriteCache.clearAddrMap()) +- http://sourceforge.net/apps/trac/bigdata/ticket/655 (SPARQL REGEX operator does not perform case-folding correctly for Unicode data) +- http://sourceforge.net/apps/trac/bigdata/ticket/656 (InFactory bug when IN args consist of a single literal) +- http://sourceforge.net/apps/trac/bigdata/ticket/647 (SIDs mode creates unnecessary hash join for GRAPH group patterns) +- http://sourceforge.net/apps/trac/bigdata/ticket/667 (Provide NanoSparqlServer initialization hook) +- http://sourceforge.net/apps/trac/bigdata/ticket/669 (Doubly nested subqueries yield no results with LIMIT) +- http://sourceforge.net/apps/trac/bigdata/ticket/675 (Flush indices in parallel during checkpoint to reduce IO latency) +- http://sourceforge.net/apps/trac/bigdata/ticket/682 (AtomicRowFilter UnsupportedOperationException) + +1.2.2: + +- http://sourceforge.net/apps/trac/bigdata/ticket/586 (RWStore immedateFree() not removing Checkpoint addresses from the historical index cache.) +- http://sourceforge.net/apps/trac/bigdata/ticket/602 (RWStore does not discard logged deletes on reset()) +- http://sourceforge.net/apps/trac/bigdata/ticket/603 (Prepare critical maintenance release as branch of 1.2.1) + +1.2.1: + +- http://sourceforge.net/apps/trac/bigdata/ticket/533 (Review materialization for inline IVs) +- http://sourceforge.net/apps/trac/bigdata/ticket/539 (NotMaterializedException with REGEX and Vocab) +- http://sourceforge.net/apps/trac/bigdata/ticket/540 (SPARQL UPDATE using NSS via index.html) +- http://sourceforge.net/apps/trac/bigdata/ticket/541 (MemoryManaged backed Journal mode) +- http://sourceforge.net/apps/trac/bigdata/ticket/546 (Index cache for Journal) +- http://sourceforge.net/apps/trac/bigdata/ticket/549 (BTree can not be cast to Name2Addr (MemStore recycler)) +- http://sourceforge.net/apps/trac/bigdata/ticket/550 (NPE in Leaf.getKey() : root cause was user error) +- http://sourceforge.net/apps/trac/bigdata/ticket/558 (SPARQL INSERT not working in same request after INSERT DATA) +- http://sourceforge.net/apps/trac/bigdata/ticket/562 (Sub-select in INSERT cause NPE in UpdateExprBuilder) +- http://sourceforge.net/apps/trac/bigdata/ticket/563 (DISTINCT ORDER BY) +- http://sourceforge.net/apps/trac/bigdata/ticket/567 (Failure to set cached value on IV results in incorrect behavior for complex UPDATE operation) +- http://sourceforge.net/apps/trac/bigdata/ticket/568 (DELETE WHERE fails with Java AssertionError) +- http://sourceforge.net/apps/trac/bigdata/ticket/569 (LOAD-CREATE-LOAD using virgin journal fails with "Graph exists" exception) +- http://sourceforge.net/apps/trac/bigdata/ticket/571 (DELETE/INSERT WHERE handling of blank nodes) +- http://sourceforge.net/apps/trac/bigdata/ticket/573 (NullPointerException when attempting to INSERT DATA containing a blank node) + +1.2.0: (*) + +- http://sourceforge.net/apps/trac/bigdata/ticket/92 (Monitoring webapp) +- http://sourceforge.net/apps/trac/bigdata/ticket/267 (Support evaluation of 3rd party operators) +- http://sourceforge.net/apps/trac/bigdata/ticket/337 (Compact and efficient movement of binding sets between nodes.) +- http://sourceforge.net/apps/trac/bigdata/ticket/433 (Cluster leaks threads under read-only index operations: DGC thread leak) +- http://sourceforge.net/apps/trac/bigdata/ticket/437 (Thread-local cache combined with unbounded thread pools causes effective memory leak: termCache memory leak & thread-local buffers) +- http://sourceforge.net/apps/trac/bigdata/ticket/438 (KeyBeforePartitionException on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/439 (Class loader problem) +- http://sourceforge.net/apps/trac/bigdata/ticket/441 (Ganglia integration) +- http://sourceforge.net/apps/trac/bigdata/ticket/443 (Logger for RWStore transaction service and recycler) +- http://sourceforge.net/apps/trac/bigdata/ticket/444 (SPARQL query can fail to notice when IRunningQuery.isDone() on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/445 (RWStore does not track tx release correctly) +- http://sourceforge.net/apps/trac/bigdata/ticket/446 (HTTP Repostory broken with bigdata 1.1.0) +- http://sourceforge.net/apps/trac/bigdata/ticket/448 (SPARQL 1.1 UPDATE) +- http://sourceforge.net/apps/trac/bigdata/ticket/449 (SPARQL 1.1 Federation extension) +- http://sourceforge.net/apps/trac/bigdata/ticket/451 (Serialization error in SIDs mode on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/454 (Global Row Store Read on Cluster uses Tx) +- http://sourceforge.net/apps/trac/bigdata/ticket/456 (IExtension implementations do point lookups on lexicon) +- http://sourceforge.net/apps/trac/bigdata/ticket/457 ("No such index" on cluster under concurrent query workload) +- http://sourceforge.net/apps/trac/bigdata/ticket/458 (Java level deadlock in DS) +- http://sourceforge.net/apps/trac/bigdata/ticket/460 (Uncaught interrupt resolving RDF terms) +- http://sourceforge.net/apps/trac/bigdata/ticket/461 (KeyAfterPartitionException / KeyBeforePartitionException on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/463 (NoSuchVocabularyItem with LUBMVocabulary for DerivedNumericsExtension) +- http://sourceforge.net/apps/trac/bigdata/ticket/464 (Query statistics do not update correctly on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/465 (Too many GRS reads on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/469 (Sail does not flush assertion buffers before query) +- http://sourceforge.net/apps/trac/bigdata/ticket/472 (acceptTaskService pool size on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/475 (Optimize serialization for query messages on cluster) +- http://sourceforge.net/apps/trac/bigdata/ticket/476 (Test suite for writeCheckpoint() and recycling for BTree/HTree) +- http://sourceforge.net/apps/trac/bigdata/ticket/478 (Cluster does not map input solution(s) across shards) +- http://sourceforge.net/apps/trac/bigdata/ticket/480 (Error releasing deferred frees using 1.0.6 against a 1.0.4 journal) +- http://sourceforge.net/apps/trac/bigdata/ticket/481 (PhysicalAddressResolutionException against 1.0.6) +- http://sourceforge.net/apps/trac/bigdata/ticket/482 (RWStore reset() should be thread-safe for concurrent readers) +- http://sourceforge.net/apps/trac/bigdata/ticket/484 (Java API for NanoSparqlServer REST API) +- http://sourceforge.net/apps/trac/bigdata/ticket/491 (AbstractTripleStore.destroy() does not clear the locator cache) +- http://sourceforge.net/apps/trac/bigdata/ticket/492 (Empty chunk in ThickChunkMessage (cluster)) +- http://sourceforge.net/apps/trac/bigdata/ticket/493 (Virtual Graphs) +- http://sourceforge.net/apps/trac/bigdata/ticket/496 (Sesame 2.6.3) +- http://sourceforge.net/apps/trac/bigdata/ticket/497 (Implement STRBEFORE, STRAFTER, and REPLACE) +- http://sourceforge.net/apps/trac/bigdata/ticket/498 (Bring bigdata RDF/XML parser up to openrdf 2.6.3.) +- http://sourceforge.net/apps/trac/bigdata/ticket/500 (SPARQL 1.1 Service Description) +- http://www.openrdf.org/issues/browse/SES-884 (Aggregation with an solution set as input should produce an empty solution as output) +- http://www.openrdf.org/issues/browse/SES-862 (Incorrect error handling for SPARQL aggregation; fix in 2.6.1) +- http://www.openrdf.org/issues/browse/SES-873 (Order the same Blank Nodes together in ORDER BY) +- http://sourceforge.net/apps/trac/bigdata/ticket/501 (SPARQL 1.1 BINDINGS are ignored) +- http://sourceforge.net/apps/trac/bigdata/ticket/503 (Bigdata2Sesame2BindingSetIterator throws QueryEvaluationException were it should throw NoSuchElementException) +- http://sourceforge.net/apps/trac/bigdata/ticket/504 (UNION with Empty Group Pattern) +- http://sourceforge.net/apps/trac/bigdata/ticket/505 (Exception when using SPARQL sort & statement identifiers) +- http://sourceforge.net/apps/trac/bigdata/ticket/506 (Load, closure and query performance in 1.1.x versus 1.0.x) +- http://sourceforge.net/apps/trac/bigdata/ticket/508 (LIMIT causes hash join utility to log errors) +- http://sourceforge.net/apps/trac/bigdata/ticket/513 (Expose the LexiconConfiguration to Function BOPs) +- http://sourceforge.net/apps/trac/bigdata/ticket/515 (Query with two "FILTER NOT EXISTS" expressions returns no results) +- http://sourceforge.net/apps/trac/bigdata/ticket/516 (REGEXBOp should cache the Pattern when it is a constant) +- http://sourceforge.net/apps/trac/bigdata/ticket/517 (Java 7 Compiler Compatibility) +- http://sourceforge.net/apps/trac/bigdata/ticket/518 (Review function bop subclass hierarchy, optimize datatype bop, etc.) +- http://sourceforge.net/apps/trac/bigdata/ticket/520 (CONSTRUCT WHERE shortcut) +- http://sourceforge.net/apps/trac/bigdata/ticket/521 (Incremental materialization of Tuple and Graph query results) +- http://sourceforge.net/apps/trac/bigdata/ticket/525 (Modify the IChangeLog interface to support multiple agents) +- http://sourceforge.net/apps/trac/bigdata/ticket/527 (Expose timestamp of LexiconRelation to function bops) +- http://sourceforge.net/apps/trac/bigdata/ticket/532 (ClassCastException during hash join (can not be cast to TermId)) +- http://sourceforge.net/apps/trac/bigdata/ticket/533 (Review materialization for inline IVs) +- http://sourceforge.net/apps/trac/bigdata/ticket/534 (BSBM BI Q5 error using MERGE JOIN) + +1.1.0 (*) + + - http://sourceforge.net/apps/trac/bigdata/ticket/23 (Lexicon joins) + - http://sourceforge.net/apps/trac/bigdata/ticket/109 (Store large literals as "blobs") + - http://sourceforge.net/apps/trac/bigdata/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) + - http://sourceforge.net/apps/trac/bigdata/ticket/203 (Implement an persistence capable hash table to support analytic query) + - http://sourceforge.net/apps/trac/bigdata/ticket/209 (AccessPath should visit binding sets rather than elements for high level query.) + - http://sourceforge.net/apps/trac/bigdata/ticket/227 (SliceOp appears to be necessary when operator plan should suffice without) + - http://sourceforge.net/apps/trac/bigdata/ticket/232 (Bottom-up evaluation semantics). + - http://sourceforge.net/apps/trac/bigdata/ticket/246 (Derived xsd numeric data types must be inlined as extension types.) + - http://sourceforge.net/apps/trac/bigdata/ticket/254 (Revisit pruning of intermediate variable bindings during query execution) + - http://sourceforge.net/apps/trac/bigdata/ticket/261 (Lift conditions out of subqueries.) + - http://sourceforge.net/apps/trac/bigdata/ticket/300 (Native ORDER BY) + - http://sourceforge.net/apps/trac/bigdata/ticket/324 (Inline predeclared URIs and namespaces in 2-3 bytes) + - http://sourceforge.net/apps/trac/bigdata/ticket/330 (NanoSparqlServer does not locate "html" resources when run from jar) + - http://sourceforge.net/apps/trac/bigdata/ticket/334 (Support inlining of unicode data in the statement indices.) + - http://sourceforge.net/apps/trac/bigdata/ticket/364 (Scalable default graph evaluation) + - http://sourceforge.net/apps/trac/bigdata/ticket/368 (Prune variable bindings during query evaluation) + - http://sourceforge.net/apps/trac/bigdata/ticket/370 (Direct translation of openrdf AST to bigdata AST) + - http://sourceforge.net/apps/trac/bigdata/ticket/373 (Fix StrBOp and other IValueExpressions) + - http://sourceforge.net/apps/trac/bigdata/ticket/377 (Optimize OPTIONALs with multiple statement patterns.) + - http://sourceforge.net/apps/trac/bigdata/ticket/380 (Native SPARQL evaluation on cluster) + - http://sourceforge.net/apps/trac/bigdata/ticket/387 (Cluster does not compute closure) + - http://sourceforge.net/apps/trac/bigdata/ticket/395 (HTree hash join performance) + - http://sourceforge.net/apps/trac/bigdata/ticket/401 (inline xsd:unsigned datatypes) + - http://sourceforge.net/apps/trac/bigdata/ticket/408 (xsd:string cast fails for non-numeric data) + - http://sourceforge.net/apps/trac/bigdata/ticket/421 (New query hints model.) + - http://sourceforge.net/apps/trac/bigdata/ticket/431 (Use of read-only tx per query defeats cache on cluster) + +1.0.3 + + - http://sourceforge.net/apps/trac/bigdata/ticket/217 (BTreeCounters does not track bytes released) + - http://sourceforge.net/apps/trac/bigdata/ticket/269 (Refactor performance counters using accessor interface) + - http://sourceforge.net/apps/trac/bigdata/ticket/329 (B+Tree should delete bloom filter when it is disabled.) + - http://sourceforge.net/apps/trac/bigdata/ticket/372 (RWStore does not prune the CommitRecordIndex) + - http://sourceforge.net/apps/trac/bigdata/ticket/375 (Persistent memory leaks (RWStore/DISK)) + - http://sourceforge.net/apps/trac/bigdata/ticket/385 (FastRDFValueCoder2: ArrayIndexOutOfBoundsException) + - http://sourceforge.net/apps/trac/bigdata/ticket/391 (Release age advanced on WORM mode journal) + - http://sourceforge.net/apps/trac/bigdata/ticket/392 (Add a DELETE by access path method to the NanoSparqlServer) + - http://sourceforge.net/apps/trac/bigdata/ticket/393 (Add "context-uri" request parameter to specify the default context for INSERT in the REST API) + - http://sourceforge.net/apps/trac/bigdata/ticket/394 (log4j configuration error message in WAR deployment) + - http://sourceforge.net/apps/trac/bigdata/ticket/399 (Add a fast range count method to the REST API) + - http://sourceforge.net/apps/trac/bigdata/ticket/422 (Support temp triple store wrapped by a BigdataSail) + - http://sourceforge.net/apps/trac/bigdata/ticket/424 (NQuads support for NanoSparqlServer) + - http://sourceforge.net/apps/trac/bigdata/ticket/425 (Bug fix to DEFAULT_RDF_FORMAT for bulk data loader in scale-out) + - http://sourceforge.net/apps/trac/bigdata/ticket/426 (Support either lockfile (procmail) and dotlockfile (liblockfile1) in scale-out) + - http://sourceforge.net/apps/trac/bigdata/ticket/427 (BigdataSail#getReadOnlyConnection() race condition with concurrent commit) + - http://sourceforge.net/apps/trac/bigdata/ticket/435 (Address is 0L) + - http://sourceforge.net/apps/trac/bigdata/ticket/436 (TestMROWTransactions failure in CI) + +1.0.2 + + - http://sourceforge.net/apps/trac/bigdata/ticket/32 (Query time expansion of (foo rdf:type rdfs:Resource) drags in SPORelation for scale-out.) + - http://sourceforge.net/apps/trac/bigdata/ticket/181 (Scale-out LUBM "how to" in wiki and build.xml are out of date.) + - http://sourceforge.net/apps/trac/bigdata/ticket/356 (Query not terminated by error.) + - http://sourceforge.net/apps/trac/bigdata/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) + - http://sourceforge.net/apps/trac/bigdata/ticket/361 (IRunningQuery not closed promptly.) + - http://sourceforge.net/apps/trac/bigdata/ticket/371 (DataLoader fails to load resources available from the classpath.) + - http://sourceforge.net/apps/trac/bigdata/ticket/376 (Support for the streaming of bigdata IBindingSets into a sparql query.) + - http://sourceforge.net/apps/trac/bigdata/ticket/378 (ClosedByInterruptException during heavy query mix.) + - http://sourceforge.net/apps/trac/bigdata/ticket/379 (NotSerializableException for SPOAccessPath.) + - http://sourceforge.net/apps/trac/bigdata/ticket/382 (Change dependencies to Apache River 2.2.0) + +1.0.1 (*) + + - http://sourceforge.net/apps/trac/bigdata/ticket/107 (Unicode clean schema names in the sparse row store). + - http://sourceforge.net/apps/trac/bigdata/ticket/124 (TermIdEncoder should use more bits for scale-out). + - http://sourceforge.net/apps/trac/bigdata/ticket/225 (OSX requires specialized performance counter collection classes). + - http://sourceforge.net/apps/trac/bigdata/ticket/348 (BigdataValueFactory.asValue() must return new instance when DummyIV is used). + - http://sourceforge.net/apps/trac/bigdata/ticket/349 (TermIdEncoder limits Journal to 2B distinct RDF Values per triple/quad store instance). + - http://sourceforge.net/apps/trac/bigdata/ticket/351 (SPO not Serializable exception in SIDS mode (scale-out)). + - http://sourceforge.net/apps/trac/bigdata/ticket/352 (ClassCastException when querying with binding-values that are not known to the database). + - http://sourceforge.net/apps/trac/bigdata/ticket/353 (UnsupportedOperatorException for some SPARQL queries). + - http://sourceforge.net/apps/trac/bigdata/ticket/355 (Query failure when comparing with non materialized value). + - http://sourceforge.net/apps/trac/bigdata/ticket/357 (RWStore reports "FixedAllocator returning null address, with freeBits".) + - http://sourceforge.net/apps/trac/bigdata/ticket/359 (NamedGraph pattern fails to bind graph variable if only one binding exists.) + - http://sourceforge.net/apps/trac/bigdata/ticket/362 (log4j - slf4j bridge.) + +For more information about bigdata(R), please see the following links: + +[1] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=Main_Page +[2] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=GettingStarted +[3] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=Roadmap +[4] http://www.bigdata.com/bigdata/docs/api/ +[5] http://sourceforge.net/projects/bigdata/ +[6] http://www.bigdata.com/blog +[7] http://www.systap.com/bigdata.htm +[8] http://sourceforge.net/projects/bigdata/files/bigdata/ +[9] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=DataMigration +[10] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=HAJournalServer +[11] http://www.bigdata.com/whitepapers/reifSPARQL.pdf +[12] http://sourceforge.net/apps/mediawiki/bigdata/index.php?title=RDF_GAS_API + +About bigdata: + +Bigdata(R) is a horizontally-scaled, general purpose storage and computing fabric for ordered data (B+Trees), designed to operate on either a single server or a cluster of commodity hardware. Bigdata(R) uses dynamically partitioned key-range shards in order to remove any realistic scaling limits - in principle, bigdata(R) may be deployed on 10s, 100s, or even thousands of machines and new capacity may be added incrementally without requiring the full reload of all data. The bigdata(R) RDF database supports RDFS and OWL Lite reasoning, high-level query (SPARQL), and datum level provenance. This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-05 18:58:41
|
Revision: 8201 http://sourceforge.net/p/bigdata/code/8201 Author: thompsonbry Date: 2014-05-05 18:58:36 +0000 (Mon, 05 May 2014) Log Message: ----------- Refactored the package for several LBS files. See #624 (HA LBS) Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournalServer.java branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/AbstractHA3LoadBalancerTestCase.java branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournalTest.java branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_GangliaLBS.java branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_NOP.java branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_RoundRobin.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/BigdataServlet.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HALoadBalancerServlet.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/ServiceScore.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java Added Paths: ----------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/DefaultHARequestURIRewriter.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHALoadBalancerPolicy.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHAPolicyLifeCycle.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHARequestURIRewriter.java Removed Paths: ------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/DefaultHARequestURIRewriter.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHALoadBalancerPolicy.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHAPolicyLifeCycle.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHARequestURIRewriter.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournalServer.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournalServer.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/java/com/bigdata/journal/jini/ha/HAJournalServer.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -104,8 +104,8 @@ import com.bigdata.rdf.sail.CreateKBTask; import com.bigdata.rdf.sail.webapp.ConfigParams; import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.NanoSparqlServer; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rwstore.RWStore; import com.bigdata.service.AbstractHATransactionService; import com.bigdata.service.jini.FakeLifeCycle; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/AbstractHA3LoadBalancerTestCase.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/AbstractHA3LoadBalancerTestCase.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/AbstractHA3LoadBalancerTestCase.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -37,9 +37,9 @@ import com.bigdata.journal.jini.ha.HAJournalServer.HAQuorumService; import com.bigdata.journal.jini.ha.HAJournalTest.HAGlueTest; import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.client.RemoteRepository; import com.bigdata.rdf.sail.webapp.client.RemoteRepository.RemoveOp; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; /** * Test suite for the HA load balancer. Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournalTest.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournalTest.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/HAJournalTest.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -102,7 +102,7 @@ import com.bigdata.rdf.sail.BigdataSailRepository; import com.bigdata.rdf.sail.BigdataSailRepositoryConnection; import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rdf.store.AbstractTripleStore; import com.bigdata.service.jini.RemoteDestroyAdmin; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_GangliaLBS.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_GangliaLBS.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_GangliaLBS.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -23,7 +23,7 @@ */ package com.bigdata.journal.jini.ha; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.lbs.policy.ganglia.GangliaLBSPolicy; /** Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_NOP.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_NOP.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_NOP.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -23,7 +23,7 @@ */ package com.bigdata.journal.jini.ha; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.lbs.policy.NOPLBSPolicy; /** Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_RoundRobin.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_RoundRobin.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-jini/src/test/com/bigdata/journal/jini/ha/TestHA3LoadBalancer_RoundRobin.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -23,7 +23,7 @@ */ package com.bigdata.journal.jini.ha; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.lbs.policy.RoundRobinLBSPolicy; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/BigdataServlet.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/BigdataServlet.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/BigdataServlet.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -44,6 +44,7 @@ import com.bigdata.journal.IIndexManager; import com.bigdata.quorum.AbstractQuorum; import com.bigdata.rdf.sail.webapp.client.IMimeTypes; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; /** * Useful glue for implementing service actions, but does not directly implement Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/DefaultHARequestURIRewriter.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/DefaultHARequestURIRewriter.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/DefaultHARequestURIRewriter.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -1,85 +0,0 @@ -/** -Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. - -Contact: - SYSTAP, LLC - 4501 Tower Road - Greensboro, NC 27410 - lic...@bi... - -This program is free software; you can redistribute it and/or modify -it under the terms of the GNU General Public License as published by -the Free Software Foundation; version 2 of the License. - -This program is distributed in the hope that it will be useful, -but WITHOUT ANY WARRANTY; without even the implied warranty of -MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -GNU General Public License for more details. - -You should have received a copy of the GNU General Public License -along with this program; if not, write to the Free Software -Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA -*/ -package com.bigdata.rdf.sail.webapp; - -import javax.servlet.ServletConfig; -import javax.servlet.ServletException; -import javax.servlet.http.HttpServletRequest; - -import com.bigdata.journal.IIndexManager; - -/** - * Default implementation. - * - * @author <a href="mailto:tho...@us...">Bryan Thompson</a> - */ -public class DefaultHARequestURIRewriter implements IHARequestURIRewriter { - - /** - * {@inheritDoc} - * <p> - * This implementation is a NOP. - */ - @Override - public void init(ServletConfig servletConfig, IIndexManager indexManager) - throws ServletException { - - } - - /** - * {@inheritDoc} - * <p> - * This implementation is a NOP. - */ - @Override - public void destroy() { - - } - - @Override - public StringBuilder rewriteURI(final boolean isLeaderRequest, - final String full_prefix, final String originalRequestURL, - final String proxyToRequestURL, final HttpServletRequest request) { - - final StringBuilder uri = new StringBuilder(proxyToRequestURL); - - if (proxyToRequestURL.endsWith("/")) - uri.setLength(uri.length() - 1); - - final String rest = originalRequestURL.substring(full_prefix.length()); - - if (!rest.startsWith("/")) - uri.append("/"); - - uri.append(rest); - - final String query = request.getQueryString(); - - if (query != null) - uri.append("?").append(query); - - return uri; - - } - -} Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HALoadBalancerServlet.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HALoadBalancerServlet.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HALoadBalancerServlet.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -42,6 +42,10 @@ import com.bigdata.journal.IIndexManager; import com.bigdata.journal.PlatformStatsPlugIn; import com.bigdata.journal.jini.ha.HAJournal; +import com.bigdata.rdf.sail.webapp.lbs.DefaultHARequestURIRewriter; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; +import com.bigdata.rdf.sail.webapp.lbs.IHAPolicyLifeCycle; +import com.bigdata.rdf.sail.webapp.lbs.IHARequestURIRewriter; import com.bigdata.rdf.sail.webapp.lbs.policy.NOPLBSPolicy; import com.bigdata.util.InnerCause; import com.bigdata.util.StackInfoReport; Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -64,6 +64,7 @@ import com.bigdata.quorum.zk.ZKQuorumClient; import com.bigdata.quorum.zk.ZKQuorumImpl; import com.bigdata.rdf.sail.webapp.StatusServlet.DigestEnum; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.zookeeper.DumpZookeeper; /** Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHALoadBalancerPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHALoadBalancerPolicy.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHALoadBalancerPolicy.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -1,86 +0,0 @@ -/** -Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. - -Contact: - SYSTAP, LLC - 4501 Tower Road - Greensboro, NC 27410 - lic...@bi... - -This program is free software; you can redistribute it and/or modify -it under the terms of the GNU General Public License as published by -the Free Software Foundation; version 2 of the License. - -This program is distributed in the hope that it will be useful, -but WITHOUT ANY WARRANTY; without even the implied warranty of -MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -GNU General Public License for more details. - -You should have received a copy of the GNU General Public License -along with this program; if not, write to the Free Software -Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA -*/ -package com.bigdata.rdf.sail.webapp; - -import java.io.IOException; - -import javax.servlet.ServletException; -import javax.servlet.http.HttpServletRequest; -import javax.servlet.http.HttpServletResponse; - -/** - * Load balancer policy interface. - * - * @author <a href="mailto:tho...@us...">Bryan Thompson</a> - * - * @see HALoadBalancerServlet - * @see <a href="http://trac.bigdata.com/ticket/624">HA Load Balancer</a> - */ -public interface IHALoadBalancerPolicy extends IHAPolicyLifeCycle { - - /** - * Invoked for each request. If the response is not committed, then it will - * be handled by the {@link HALoadBalancerServlet}. - * - * @param isLeaderRequest - * <code>true</code> iff this request must be directed to the - * leaeder and <code>false</code> iff this request may be load - * balanced over the joined services. UPDATEs MUST be handled by - * the leader. Read requests can be handled by any service that - * is joined with the met quorum. - * @param request - * The request. - * @param response - * The response. - * - * @return <code>true</code> iff the request was handled. - */ - boolean service(final boolean isLeaderRequest, - final HttpServletRequest request, final HttpServletResponse response) - throws ServletException, IOException; - - /** - * Return the URL to which a non-idempotent request will be proxied. - * - * @param req - * The request. - * - * @return The proxyTo URL -or- <code>null</code> if we could not find a - * service to which we could proxy this request. - */ - String getLeaderURL(HttpServletRequest req); - - /** - * Return the URL to which a <strong>read-only</strong> request will be - * proxied. The returned URL must include the protocol, hostname and port - * (if a non-default port will be used) as well as the target request path. - * - * @param req - * The request. - * - * @return The proxyTo URL -or- <code>null</code> if we could not find a - * service to which we could proxy this request. - */ - String getReaderURL(HttpServletRequest req); - -} Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHAPolicyLifeCycle.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHAPolicyLifeCycle.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHAPolicyLifeCycle.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -1,47 +0,0 @@ -/** -Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. - -Contact: - SYSTAP, LLC - 4501 Tower Road - Greensboro, NC 27410 - lic...@bi... - -This program is free software; you can redistribute it and/or modify -it under the terms of the GNU General Public License as published by -the Free Software Foundation; version 2 of the License. - -This program is distributed in the hope that it will be useful, -but WITHOUT ANY WARRANTY; without even the implied warranty of -MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -GNU General Public License for more details. - -You should have received a copy of the GNU General Public License -along with this program; if not, write to the Free Software -Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA -*/ -package com.bigdata.rdf.sail.webapp; - -import javax.servlet.ServletConfig; -import javax.servlet.ServletException; - -import com.bigdata.journal.IIndexManager; - -public interface IHAPolicyLifeCycle { - - /** - * Initialize the policy. - * - * @param servletConfig - * @param indexManager - */ - void init(ServletConfig servletConfig, IIndexManager indexManager) - throws ServletException; - - /** - * Destroy the policy (stop any asynchronous processing, release any - * resources). - */ - void destroy(); - -} Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHARequestURIRewriter.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHARequestURIRewriter.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHARequestURIRewriter.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -1,95 +0,0 @@ -/** -Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. - -Contact: - SYSTAP, LLC - 4501 Tower Road - Greensboro, NC 27410 - lic...@bi... - -This program is free software; you can redistribute it and/or modify -it under the terms of the GNU General Public License as published by -the Free Software Foundation; version 2 of the License. - -This program is distributed in the hope that it will be useful, -but WITHOUT ANY WARRANTY; without even the implied warranty of -MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -GNU General Public License for more details. - -You should have received a copy of the GNU General Public License -along with this program; if not, write to the Free Software -Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA -*/ -package com.bigdata.rdf.sail.webapp; - -import javax.servlet.http.HttpServletRequest; - -import com.bigdata.rdf.sail.webapp.lbs.ServiceScore; - -/** - * Interface for rewriting the Request-URI once the load balancer has determined - * the target host and service to which the request will be proxied. - * - * @author <a href="mailto:tho...@us...">Bryan Thompson</a> - */ -public interface IHARequestURIRewriter extends IHAPolicyLifeCycle { - - /** - * Rewrite the <code>originalRequestURI</code> into a <a href= - * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" - * >Request-URL</a> for the web application whose servlet context root is - * given by the <code>proxyToRequestURI</code>. - * <p> - * Note: The <code>proxyToRequestURI</code> is include the protocol, host, - * port, and servlet context path for the target service. It DOES NOT - * include any information from the original request. The purpose of this - * method is to modify the <code>originalRequestURI</code> in order to - * obtain a fully qualified RequestURI for the service to which the request - * will be proxied. For example: - * - * <pre> - * full_prefix: /bigdata/LBS/leader - * -or- full_prefix: /bigdata/LBS/read - * originalRequestURI: http://ha1.example.com:8090/bigdata/LBS/read/sparql - * proxyToRequestURI: http://ha3.example.com:8090/bigdata/ - * return: http://ha2.example.com:8090/bigdata/LBS/read/sparql - * </pre> - * <p> - * Note: this method is only invoked if we will proxy to another service. - * Therefore, the <code>proxyToRequestURI</code> is never <code>null</code>. - * - * @param isLeaderRequest - * <code>true</code> iff the request is directed to the leader. - * @param full_prefix - * The path prefix in the <code>originalRequestURI</code> that - * which corresponds to the {@link HALoadBalancerServlet} and - * which must be removed if the request is to be forwarded to a - * local service. - * @param originalRequestURI - * The original <a href= - * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" - * >Request-URL</a> from the <a - * href="http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1" - * >HTTP Request-Line</a> - * @param proxyToRequestURI - * The RequestURI for the root of the web application for the - * target service and never <code>null</code>. - * @param request - * The original request. - * - * @return The fully qualified <a href= - * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" - * >Request-URL</a> that will be used to proxy the http request to - * the service identified by the <code>proxyToRequestURI</code> - * - * @see ServiceScore#getRequestURI() - */ - public StringBuilder rewriteURI(// - boolean isLeaderRequest, - String full_prefix,// - String originalRequestURI, // - String proxyToRequestURI,// - HttpServletRequest request// - ); - -} Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -51,7 +51,6 @@ import com.bigdata.quorum.QuorumListener; import com.bigdata.rdf.sail.webapp.BigdataServlet; import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; /** * Abstract base class establishes a listener for quorum events, tracks the Copied: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/DefaultHARequestURIRewriter.java (from rev 8166, branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/DefaultHARequestURIRewriter.java) =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/DefaultHARequestURIRewriter.java (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/DefaultHARequestURIRewriter.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -0,0 +1,85 @@ +/** +Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. + +Contact: + SYSTAP, LLC + 4501 Tower Road + Greensboro, NC 27410 + lic...@bi... + +This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; version 2 of the License. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program; if not, write to the Free Software +Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +*/ +package com.bigdata.rdf.sail.webapp.lbs; + +import javax.servlet.ServletConfig; +import javax.servlet.ServletException; +import javax.servlet.http.HttpServletRequest; + +import com.bigdata.journal.IIndexManager; + +/** + * Default implementation. + * + * @author <a href="mailto:tho...@us...">Bryan Thompson</a> + */ +public class DefaultHARequestURIRewriter implements IHARequestURIRewriter { + + /** + * {@inheritDoc} + * <p> + * This implementation is a NOP. + */ + @Override + public void init(ServletConfig servletConfig, IIndexManager indexManager) + throws ServletException { + + } + + /** + * {@inheritDoc} + * <p> + * This implementation is a NOP. + */ + @Override + public void destroy() { + + } + + @Override + public StringBuilder rewriteURI(final boolean isLeaderRequest, + final String full_prefix, final String originalRequestURL, + final String proxyToRequestURL, final HttpServletRequest request) { + + final StringBuilder uri = new StringBuilder(proxyToRequestURL); + + if (proxyToRequestURL.endsWith("/")) + uri.setLength(uri.length() - 1); + + final String rest = originalRequestURL.substring(full_prefix.length()); + + if (!rest.startsWith("/")) + uri.append("/"); + + uri.append(rest); + + final String query = request.getQueryString(); + + if (query != null) + uri.append("?").append(query); + + return uri; + + } + +} Copied: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHALoadBalancerPolicy.java (from rev 8166, branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHALoadBalancerPolicy.java) =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHALoadBalancerPolicy.java (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHALoadBalancerPolicy.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -0,0 +1,88 @@ +/** +Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. + +Contact: + SYSTAP, LLC + 4501 Tower Road + Greensboro, NC 27410 + lic...@bi... + +This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; version 2 of the License. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program; if not, write to the Free Software +Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +*/ +package com.bigdata.rdf.sail.webapp.lbs; + +import java.io.IOException; + +import javax.servlet.ServletException; +import javax.servlet.http.HttpServletRequest; +import javax.servlet.http.HttpServletResponse; + +import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; + +/** + * Load balancer policy interface. + * + * @author <a href="mailto:tho...@us...">Bryan Thompson</a> + * + * @see HALoadBalancerServlet + * @see <a href="http://trac.bigdata.com/ticket/624">HA Load Balancer</a> + */ +public interface IHALoadBalancerPolicy extends IHAPolicyLifeCycle { + + /** + * Invoked for each request. If the response is not committed, then it will + * be handled by the {@link HALoadBalancerServlet}. + * + * @param isLeaderRequest + * <code>true</code> iff this request must be directed to the + * leaeder and <code>false</code> iff this request may be load + * balanced over the joined services. UPDATEs MUST be handled by + * the leader. Read requests can be handled by any service that + * is joined with the met quorum. + * @param request + * The request. + * @param response + * The response. + * + * @return <code>true</code> iff the request was handled. + */ + boolean service(final boolean isLeaderRequest, + final HttpServletRequest request, final HttpServletResponse response) + throws ServletException, IOException; + + /** + * Return the URL to which a non-idempotent request will be proxied. + * + * @param req + * The request. + * + * @return The proxyTo URL -or- <code>null</code> if we could not find a + * service to which we could proxy this request. + */ + String getLeaderURL(HttpServletRequest req); + + /** + * Return the URL to which a <strong>read-only</strong> request will be + * proxied. The returned URL must include the protocol, hostname and port + * (if a non-default port will be used) as well as the target request path. + * + * @param req + * The request. + * + * @return The proxyTo URL -or- <code>null</code> if we could not find a + * service to which we could proxy this request. + */ + String getReaderURL(HttpServletRequest req); + +} Copied: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHAPolicyLifeCycle.java (from rev 8166, branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHAPolicyLifeCycle.java) =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHAPolicyLifeCycle.java (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHAPolicyLifeCycle.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -0,0 +1,47 @@ +/** +Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. + +Contact: + SYSTAP, LLC + 4501 Tower Road + Greensboro, NC 27410 + lic...@bi... + +This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; version 2 of the License. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program; if not, write to the Free Software +Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +*/ +package com.bigdata.rdf.sail.webapp.lbs; + +import javax.servlet.ServletConfig; +import javax.servlet.ServletException; + +import com.bigdata.journal.IIndexManager; + +public interface IHAPolicyLifeCycle { + + /** + * Initialize the policy. + * + * @param servletConfig + * @param indexManager + */ + void init(ServletConfig servletConfig, IIndexManager indexManager) + throws ServletException; + + /** + * Destroy the policy (stop any asynchronous processing, release any + * resources). + */ + void destroy(); + +} Copied: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHARequestURIRewriter.java (from rev 8166, branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/IHARequestURIRewriter.java) =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHARequestURIRewriter.java (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/IHARequestURIRewriter.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -0,0 +1,95 @@ +/** +Copyright (C) SYSTAP, LLC 2006-2007. All rights reserved. + +Contact: + SYSTAP, LLC + 4501 Tower Road + Greensboro, NC 27410 + lic...@bi... + +This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; version 2 of the License. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program; if not, write to the Free Software +Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +*/ +package com.bigdata.rdf.sail.webapp.lbs; + +import javax.servlet.http.HttpServletRequest; + +import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; + +/** + * Interface for rewriting the Request-URI once the load balancer has determined + * the target host and service to which the request will be proxied. + * + * @author <a href="mailto:tho...@us...">Bryan Thompson</a> + */ +public interface IHARequestURIRewriter extends IHAPolicyLifeCycle { + + /** + * Rewrite the <code>originalRequestURI</code> into a <a href= + * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" + * >Request-URL</a> for the web application whose servlet context root is + * given by the <code>proxyToRequestURI</code>. + * <p> + * Note: The <code>proxyToRequestURI</code> is include the protocol, host, + * port, and servlet context path for the target service. It DOES NOT + * include any information from the original request. The purpose of this + * method is to modify the <code>originalRequestURI</code> in order to + * obtain a fully qualified RequestURI for the service to which the request + * will be proxied. For example: + * + * <pre> + * full_prefix: /bigdata/LBS/leader + * -or- full_prefix: /bigdata/LBS/read + * originalRequestURI: http://ha1.example.com:8090/bigdata/LBS/read/sparql + * proxyToRequestURI: http://ha3.example.com:8090/bigdata/ + * return: http://ha2.example.com:8090/bigdata/LBS/read/sparql + * </pre> + * <p> + * Note: this method is only invoked if we will proxy to another service. + * Therefore, the <code>proxyToRequestURI</code> is never <code>null</code>. + * + * @param isLeaderRequest + * <code>true</code> iff the request is directed to the leader. + * @param full_prefix + * The path prefix in the <code>originalRequestURI</code> that + * which corresponds to the {@link HALoadBalancerServlet} and + * which must be removed if the request is to be forwarded to a + * local service. + * @param originalRequestURI + * The original <a href= + * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" + * >Request-URL</a> from the <a + * href="http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1" + * >HTTP Request-Line</a> + * @param proxyToRequestURI + * The RequestURI for the root of the web application for the + * target service and never <code>null</code>. + * @param request + * The original request. + * + * @return The fully qualified <a href= + * "http://www.w3.org/Protocols/rfc2616/rfc2616-sec5.html#sec5.1.2" + * >Request-URL</a> that will be used to proxy the http request to + * the service identified by the <code>proxyToRequestURI</code> + * + * @see ServiceScore#getRequestURI() + */ + public StringBuilder rewriteURI(// + boolean isLeaderRequest, + String full_prefix,// + String originalRequestURI, // + String proxyToRequestURI,// + HttpServletRequest request// + ); + +} Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/ServiceScore.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/ServiceScore.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/ServiceScore.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -31,7 +31,6 @@ import com.bigdata.journal.IIndexManager; import com.bigdata.journal.jini.ha.HAJournal; import com.bigdata.quorum.Quorum; -import com.bigdata.rdf.sail.webapp.IHARequestURIRewriter; /** * Helper class caches metadata about an {@link HAGlue} service. Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-05 17:17:10 UTC (rev 8200) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-05 18:58:36 UTC (rev 8201) @@ -52,9 +52,9 @@ import com.bigdata.journal.jini.ha.HAJournal; import com.bigdata.journal.jini.ha.HAJournalServer; import com.bigdata.rdf.sail.webapp.HALoadBalancerServlet; -import com.bigdata.rdf.sail.webapp.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.lbs.AbstractLBSPolicy; import com.bigdata.rdf.sail.webapp.lbs.HostScore; +import com.bigdata.rdf.sail.webapp.lbs.IHALoadBalancerPolicy; import com.bigdata.rdf.sail.webapp.lbs.ServiceScore; import com.bigdata.util.InnerCause; This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tob...@us...> - 2014-05-05 17:17:13
|
Revision: 8200 http://sourceforge.net/p/bigdata/code/8200 Author: tobycraig Date: 2014-05-05 17:17:10 +0000 (Mon, 05 May 2014) Log Message: ----------- #891 - Added GAS namespace to shortcuts Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-05 17:00:10 UTC (rev 8199) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-05 17:17:10 UTC (rev 8200) @@ -198,6 +198,7 @@ 'owl': 'http://www.w3.org/2002/07/owl#', 'bd': 'http://www.bigdata.com/rdf#', 'bds': 'http://www.bigdata.com/rdf/search#', + 'gas': 'http://www.bigdata.com/rdf/gas#', 'foaf': 'http://xmlns.com/foaf/0.1/', 'hint': 'http://www.bigdata.com/queryHints#', 'dc': 'http://purl.org/dc/elements/1.1/', This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tob...@us...> - 2014-05-05 17:00:14
|
Revision: 8199 http://sourceforge.net/p/bigdata/code/8199 Author: tobycraig Date: 2014-05-05 17:00:10 +0000 (Mon, 05 May 2014) Log Message: ----------- #906 - Fixed namespace shortcut buttons for namespaces that don't end with # Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-05 16:24:41 UTC (rev 8198) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/js/workbench.js 2014-05-05 17:00:10 UTC (rev 8199) @@ -206,7 +206,10 @@ $('.namespace-shortcuts').html('<ul>'); for(var ns in NAMESPACE_SHORTCUTS) { - $('.namespace-shortcuts ul').append('<li data-ns="prefix ' + ns + ': <' + NAMESPACE_SHORTCUTS[ns] + '>">' + ns.toUpperCase() + '</li>'); + // cannot use data-ns attribute on li, as jQuery mangles namespaces that don't end with #, adding </li> to them + var li = $('<li>' + ns.toUpperCase() + '</li>'); + li.data('ns', 'prefix ' + ns + ': <' + NAMESPACE_SHORTCUTS[ns] + '>'); + li.appendTo('.namespace-shortcuts ul'); } $('.namespace-shortcuts li').click(function() { This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <mrp...@us...> - 2014-05-05 16:24:44
|
Revision: 8198 http://sourceforge.net/p/bigdata/code/8198 Author: mrpersonick Date: 2014-05-05 16:24:41 +0000 (Mon, 05 May 2014) Log Message: ----------- Ticket #907: bigdata quick start. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/build.xml Modified: branches/BIGDATA_RELEASE_1_3_0/build.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/build.xml 2014-05-05 15:38:22 UTC (rev 8197) +++ branches/BIGDATA_RELEASE_1_3_0/build.xml 2014-05-05 16:24:41 UTC (rev 8198) @@ -2464,6 +2464,17 @@ </java> </target> + <target name="start" depends="compile" description="Start the Bigdata Server."> + <java classname="com.bigdata.rdf.sail.webapp.NanoSparqlServer" failonerror="true" fork="true" logerror="true"> + <classpath refid="runtime.classpath" /> + <jvmarg value="-server"/> + <jvmarg value="-Xmx1G"/> + <jvmarg value="-Dlog4j.configuration=bigdata-war/src/WEB-INF/classes/log4j.properties"/> + <arg value="9999"/> + <arg value="kb"/> + <arg value="bigdata-war/src/WEB-INF/RWStore.properties"/> + </java> + </target> </project> This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tob...@us...> - 2014-05-05 15:38:24
|
Revision: 8197 http://sourceforge.net/p/bigdata/code/8197 Author: tobycraig Date: 2014-05-05 15:38:22 +0000 (Mon, 05 May 2014) Log Message: ----------- Resized logo Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/images/logo.png Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/html/images/logo.png =================================================================== (Binary files differ) This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <mrp...@us...> - 2014-05-05 15:08:40
|
Revision: 8196 http://sourceforge.net/p/bigdata/code/8196 Author: mrpersonick Date: 2014-05-05 15:08:22 +0000 (Mon, 05 May 2014) Log Message: ----------- got rid of IDoNotJoinService Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/controller/ServiceCallJoin.java Removed Paths: ------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/service/IDoNotJoinService.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/controller/ServiceCallJoin.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/controller/ServiceCallJoin.java 2014-05-05 12:29:04 UTC (rev 8195) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/controller/ServiceCallJoin.java 2014-05-05 15:08:22 UTC (rev 8196) @@ -58,7 +58,6 @@ import com.bigdata.rdf.model.BigdataURI; import com.bigdata.rdf.sparql.ast.service.BigdataServiceCall; import com.bigdata.rdf.sparql.ast.service.ExternalServiceCall; -import com.bigdata.rdf.sparql.ast.service.IDoNotJoinService; import com.bigdata.rdf.sparql.ast.service.RemoteServiceCall; import com.bigdata.rdf.sparql.ast.service.ServiceCall; import com.bigdata.rdf.sparql.ast.service.ServiceCallUtility; @@ -586,52 +585,6 @@ : new UnsyncLocalOutputBuffer<IBindingSet>( op.getChunkCapacity(), sink2); - if (serviceCall instanceof IDoNotJoinService) { - - // The iterator draining the subquery - ICloseableIterator<IBindingSet[]> serviceSolutionItr = null; - try { - - /* - * Invoke the service. - * - * Note: Returns [null] IFF SILENT and SERVICE ERROR. - */ - - serviceSolutionItr = doServiceCall(serviceCall, chunk); - - if (serviceSolutionItr != null) { - - while (serviceSolutionItr.hasNext()) { - - final IBindingSet[] bsets = - serviceSolutionItr.next(); - - for (IBindingSet bs : bsets) { - - unsyncBuffer.add(bs); - - } - - } - - } - - } finally { - - // ensure the service call iterator is closed. - if (serviceSolutionItr != null) - serviceSolutionItr.close(); - - } - - unsyncBuffer.flush(); - - // done. - return null; - - } - final JVMHashJoinUtility state = new JVMHashJoinUtility(op, silent ? JoinTypeEnum.Optional : JoinTypeEnum.Normal ); Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/service/IDoNotJoinService.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/service/IDoNotJoinService.java 2014-05-05 12:29:04 UTC (rev 8195) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/service/IDoNotJoinService.java 2014-05-05 15:08:22 UTC (rev 8196) @@ -1,35 +0,0 @@ -/** - -Copyright (C) SYSTAP, LLC 2006-2014. All rights reserved. - -Contact: - SYSTAP, LLC - 4501 Tower Road - Greensboro, NC 27410 - lic...@bi... - -This program is free software; you can redistribute it and/or modify -it under the terms of the GNU General Public License as published by -the Free Software Foundation; version 2 of the License. - -This program is distributed in the hope that it will be useful, -but WITHOUT ANY WARRANTY; without even the implied warranty of -MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the -GNU General Public License for more details. - -You should have received a copy of the GNU General Public License -along with this program; if not, write to the Free Software -Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA -*/ -package com.bigdata.rdf.sparql.ast.service; - -/** - * Service calls can implement this interface and they will not be routed - * through a hash join in the query plan. They will be responsible for their - * own join internally. - * - * @author mikepersonick - */ -public interface IDoNotJoinService { - -} This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-05 12:29:08
|
Revision: 8195 http://sourceforge.net/p/bigdata/code/8195 Author: thompsonbry Date: 2014-05-05 12:29:04 +0000 (Mon, 05 May 2014) Log Message: ----------- Modified the GangliaLBSPolicy to explicitly check for a null hostname (can be caused by a failed RMI). Modified the HA status page to include the proxy object for the RMI interface for the local service. I want to use this to diagnose situations where one service as a proxy for another but the proxy that is has does not respond and is, hence, probably dead. Note that moving into an error state on that service whose exported proxy is not responding is not enough to get the service to become responsive. This is probably because the error state does not unexport the proxy. However, it does not explain why/how a bad proxy got out there in the first place. See #624 (HA LBS) Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java 2014-05-05 11:59:15 UTC (rev 8194) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/HAStatusServletUtil.java 2014-05-05 12:29:04 UTC (rev 8195) @@ -242,6 +242,9 @@ .close(); p.text("Service: path=" + quorumService.getServiceDir()) .node("br").close(); + p.text("Service: proxy=" + + journal.getHAJournalServer().getProxy()) + .node("br").close(); } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-05 11:59:15 UTC (rev 8194) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-05 12:29:04 UTC (rev 8195) @@ -625,8 +625,12 @@ for (ServiceScore serviceScore : serviceScores) { if (serviceScore == null) // should never be null. continue; - tmp.add(serviceScore.getHostname()); + final String hostname = serviceScore.getHostname(); + if (hostname == null) // should never be null. + continue; + tmp.add(hostname); } + // dense array of hosts names for services. hosts = tmp.toArray(new String[tmp.size()]); } This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-05 11:59:19
|
Revision: 8194 http://sourceforge.net/p/bigdata/code/8194 Author: thompsonbry Date: 2014-05-05 11:59:15 +0000 (Mon, 05 May 2014) Log Message: ----------- Published new version of bigdata-ganglia (1.0.2) with new APIs for GangliaService that are used by the GangliaLBSPolicy. The artifact has been pushed to the systap maven repository. See #624 (HA LBS) Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/build.properties branches/BIGDATA_RELEASE_1_3_0/build.properties branches/BIGDATA_RELEASE_1_3_0/pom.xml Added Paths: ----------- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/src/releases/bigdata-ganglia-1.0.2.txt Removed Paths: ------------- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.1.jar Deleted: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.1.jar =================================================================== (Binary files differ) Added: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar =================================================================== (Binary files differ) Index: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar 2014-05-05 01:10:28 UTC (rev 8193) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar 2014-05-05 11:59:15 UTC (rev 8194) Property changes on: branches/BIGDATA_RELEASE_1_3_0/bigdata/lib/bigdata-ganglia-1.0.2.jar ___________________________________________________________________ Added: svn:mime-type ## -0,0 +1 ## +application/octet-stream \ No newline at end of property Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/build.properties =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/build.properties 2014-05-05 01:10:28 UTC (rev 8193) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/build.properties 2014-05-05 11:59:15 UTC (rev 8194) @@ -38,7 +38,7 @@ release.dir=ant-release # The build version. -build.ver=1.0.1 +build.ver=1.0.2 # Set true to do a snapshot build. This changes the value of ${version} to # include the date. Added: branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/src/releases/bigdata-ganglia-1.0.2.txt =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/src/releases/bigdata-ganglia-1.0.2.txt (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-ganglia/src/releases/bigdata-ganglia-1.0.2.txt 2014-05-05 11:59:15 UTC (rev 8194) @@ -0,0 +1,55 @@ +This library provides a pure Java embedded peer for Ganglia. The GangliaService +both listens and reports metrics. This means that it is capable of provide load +balanced reported from soft state and can even be used as a substitute for gmond +on operating systems (such as Windows) to which gmond has not been ported. + +The main entry point is GangliaService. It is trivial to setup with defaults and +you can easily register your own metrics collection classes to report out on your +application. + +GangliaServer service = new GangliaService("MyService"); +// Register to collect metrics. +service.addMetricCollector(new MyMetricsCollector()); +// Join the ganglia network; Start collecting and reporting metrics. +service.run(); + +The following will return the default load balanced report, which contains +exactly the same information that you would get from gstat -a. You can also use +an alternative method signature to get a report based on your own list of metrics +and/or have the report sorted by the metric (or even a synthetic metric) of your +choice. + +IHostReport[] hostReport = service.getHostReport(); + +Have fun! + +Change log: + +1.0.2: + +- Minor API additions for GangliaService. + +1.0.1: + +- Added utility class for parsing a string into an array of host addresses. +- GangliaListener was ignoring interrupts. + +-------------------------- + + Copyright (C) SYSTAP, LLC 2006-2012. All rights reserved. + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. + +---- +This product includes software developed by The Apache Software Foundation (http://www.apache.org/). +License: http://www.apache.org/licenses/LICENSE-2.0 Modified: branches/BIGDATA_RELEASE_1_3_0/build.properties =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/build.properties 2014-05-05 01:10:28 UTC (rev 8193) +++ branches/BIGDATA_RELEASE_1_3_0/build.properties 2014-05-05 11:59:15 UTC (rev 8194) @@ -69,7 +69,7 @@ fastutil.version=5.1.5 dsiutils.version=1.0.6-020610 lgplutils.version=1.0.7-270114 -ganglia-version=1.0.1 +ganglia-version=1.0.2 gas-version=0.1.0 jackson-version=2.2.3 Modified: branches/BIGDATA_RELEASE_1_3_0/pom.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/pom.xml 2014-05-05 01:10:28 UTC (rev 8193) +++ branches/BIGDATA_RELEASE_1_3_0/pom.xml 2014-05-05 11:59:15 UTC (rev 8194) @@ -97,7 +97,7 @@ <fastutil.version>5.1.5</fastutil.version> <dsiutils.version>1.0.6-020610</dsiutils.version> <lgplutils.version>1.0.7-270114</lgplutils.version> - <bigdata.ganglia.version>1.0.1</bigdata.ganglia.version> + <bigdata.ganglia.version>1.0.2</bigdata.ganglia.version> <jackson.version>2.2.3</jackson.version> </properties> <!-- TODO Can we declare the versions of the dependencies here as @@ -364,11 +364,11 @@ mvn deploy:deploy-file \ -DgroupId=com.bigdata \ -DartifactId=bigdata-ganglia \ - -Dversion=1.0.1 \ + -Dversion=1.0.2 \ -Dpackaging=jar \ -DrepositoryId=bigdata.releases \ -Durl=scpexe://www.systap.com/srv/www/htdocs/systap.com/maven/releases/ \ - -Dfile=bigdata/lib/bigdata-ganglia-1.0.1.jar + -Dfile=bigdata/lib/bigdata-ganglia-1.0.2.jar --> <groupId>com.bigdata</groupId> <artifactId>bigdata-ganglia</artifactId> This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <jer...@us...> - 2014-05-05 01:10:31
|
Revision: 8193 http://sourceforge.net/p/bigdata/code/8193 Author: jeremy_carroll Date: 2014-05-05 01:10:28 +0000 (Mon, 05 May 2014) Log Message: ----------- File missed from last commit Added Paths: ----------- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestASTPropertyPathOptimizer.java Added: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestASTPropertyPathOptimizer.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestASTPropertyPathOptimizer.java (rev 0) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestASTPropertyPathOptimizer.java 2014-05-05 01:10:28 UTC (rev 8193) @@ -0,0 +1,159 @@ +/** + +Copyright (C) SYSTAP, LLC 2006-2011. All rights reserved. + +Contact: + SYSTAP, LLC + 4501 Tower Road + Greensboro, NC 27410 + lic...@bi... + +This program is free software; you can redistribute it and/or modify +it under the terms of the GNU General Public License as published by +the Free Software Foundation; version 2 of the License. + +This program is distributed in the hope that it will be useful, +but WITHOUT ANY WARRANTY; without even the implied warranty of +MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +GNU General Public License for more details. + +You should have received a copy of the GNU General Public License +along with this program; if not, write to the Free Software +Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA +*/ +/* + * Created on Aug 29, 2011 + */ + +package com.bigdata.rdf.sparql.ast.optimizers; + +import static com.bigdata.rdf.sparql.ast.optimizers.AbstractOptimizerTestCase.HelperFlag.ZERO_OR_MORE; + +import org.openrdf.model.impl.URIImpl; +import org.openrdf.model.vocabulary.RDF; + +import com.bigdata.bop.IBindingSet; +import com.bigdata.rdf.internal.IV; +import com.bigdata.rdf.sparql.ast.ArbitraryLengthPathNode; +import com.bigdata.rdf.sparql.ast.ConstantNode; +import com.bigdata.rdf.sparql.ast.FilterNode; +import com.bigdata.rdf.sparql.ast.FunctionNode; +import com.bigdata.rdf.sparql.ast.FunctionRegistry; +import com.bigdata.rdf.sparql.ast.IQueryNode; +import com.bigdata.rdf.sparql.ast.JoinGroupNode; +import com.bigdata.rdf.sparql.ast.ProjectionNode; +import com.bigdata.rdf.sparql.ast.QueryRoot; +import com.bigdata.rdf.sparql.ast.QueryType; +import com.bigdata.rdf.sparql.ast.StatementPatternNode; +import com.bigdata.rdf.sparql.ast.UnionNode; +import com.bigdata.rdf.sparql.ast.ValueExpressionNode; +import com.bigdata.rdf.sparql.ast.VarNode; + +/** + * Test suite for {@link ASTUnionFiltersOptimizer}. + * + * @author <a href="mailto:mrp...@us...">Mike Personick</a> + * @version $Id: TestASTEmptyGroupOptimizer.java 5302 2011-10-07 14:28:03Z + * thompsonbry $ + */ +public class TestASTPropertyPathOptimizer extends AbstractOptimizerTestCase { + + /** + * + */ + public TestASTPropertyPathOptimizer() { + } + + /** + * @param name + */ + public TestASTPropertyPathOptimizer(String name) { + super(name); + } + + + @Override + IASTOptimizer newOptimizer() { + return new ASTPropertyPathOptimizerInTest(); + } + + /** + * This is (nearly) the same as {@link TestALPPinTrac773#testSimpleALPP() + */ + public void test_basic_star() { + new Helper(){{ + + given = select( varNode(x), + where ( + joinGroupNode( + propertyPathNode(varNode(x),"c*", constantNode(b)) + ) + ) ); + + + expected = select( varNode(x), + where ( + joinGroupNode( + arbitartyLengthPropertyPath(varNode(x), constantNode(b), ZERO_OR_MORE, + joinGroupNode( + statementPatternNode(leftVar(), constantNode(c), rightVar()) + ) ) ) + ) ); + }}.test(); + } + + public void test_filter_star() { + new Helper(){{ + + given = select( varNode(x), + where ( + filter( + exists(varNode(y), + joinGroupNode( + propertyPathNode(varNode(x),"c*", constantNode(b)) + ) ) ) + ) ); + + expected = select( varNode(x), + where ( + filter( + exists(varNode(y), + joinGroupNode( + arbitartyLengthPropertyPath(varNode(x), constantNode(b), ZERO_OR_MORE, + joinGroupNode( + statementPatternNode(leftVar(), constantNode(c), rightVar()) + ) ) ) ) ) + ) ); + }}.test(); + } + public void test_filter_or_star() { + new Helper(){{ + + given = select( varNode(x), + where ( + filter( + or ( + constantNode(a), + exists(varNode(y), + joinGroupNode( + propertyPathNode(varNode(x),"c*", constantNode(b)) + ) ) ) + ) + ) ); + + expected = select( varNode(x), + where ( + filter( + or ( + constantNode(a), + exists(varNode(y), + joinGroupNode( + arbitartyLengthPropertyPath(varNode(x), constantNode(b), ZERO_OR_MORE, + joinGroupNode( + statementPatternNode(leftVar(), constantNode(c), rightVar()) + ) ) ) ) ) ) + ) ); + }}.test(); + } + +} This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <jer...@us...> - 2014-05-04 23:19:43
|
Revision: 8192 http://sourceforge.net/p/bigdata/code/8192 Author: jeremy_carroll Date: 2014-05-04 23:19:37 +0000 (Sun, 04 May 2014) Log Message: ----------- Test and fix for trac904 FILTER EXISTS || TRUE. AbstractJoinGroupOptimizer needs to recurse into the filter expression looking for EXISTS Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/optimizers/AbstractJoinGroupOptimizer.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/AbstractOptimizerTestCase.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestALPPinTrac773.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestAll.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/optimizers/AbstractJoinGroupOptimizer.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/optimizers/AbstractJoinGroupOptimizer.java 2014-05-04 21:33:58 UTC (rev 8191) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/optimizers/AbstractJoinGroupOptimizer.java 2014-05-04 23:19:37 UTC (rev 8192) @@ -27,6 +27,8 @@ package com.bigdata.rdf.sparql.ast.optimizers; +import java.util.Iterator; + import com.bigdata.bop.BOp; import com.bigdata.bop.IBindingSet; import com.bigdata.bop.bindingSet.ListBindingSet; @@ -188,20 +190,7 @@ final IValueExpressionNode ve = filter.getValueExpressionNode(); - if (ve instanceof SubqueryFunctionNodeBase) { - - final SubqueryFunctionNodeBase subqueryFunction = (SubqueryFunctionNodeBase) ve; - - final GraphPatternGroup<IGroupMemberNode> graphPattern = subqueryFunction - .getGraphPattern(); - - if (graphPattern != null) { - - optimize(ctx, sa, bSets, graphPattern); - - } - - } + optimize(ctx, sa, bSets, ve); } else if (child instanceof ArbitraryLengthPathNode) { @@ -274,6 +263,35 @@ } } + + private void optimize(final AST2BOpContext ctx, final StaticAnalysis sa, + final IBindingSet[] bSets, final IValueExpressionNode ve) { + if (ve instanceof SubqueryFunctionNodeBase) { + + final SubqueryFunctionNodeBase subqueryFunction = (SubqueryFunctionNodeBase) ve; + + final GraphPatternGroup<IGroupMemberNode> graphPattern = subqueryFunction + .getGraphPattern(); + + if (graphPattern != null) { + + optimize(ctx, sa, bSets, graphPattern); + + } + + } else { + Iterator<BOp> it = ((BOp)ve).argIterator(); + while (it.hasNext()) { + + BOp b = it.next(); + if (b instanceof IValueExpressionNode) { + + optimize(ctx, sa, bSets, (IValueExpressionNode)b); + + } + } + } + } /** * Subclasses can do the work of optimizing a join group here. Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/AbstractOptimizerTestCase.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/AbstractOptimizerTestCase.java 2014-05-04 21:33:58 UTC (rev 8191) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/AbstractOptimizerTestCase.java 2014-05-04 23:19:37 UTC (rev 8192) @@ -488,6 +488,17 @@ } } + protected static final class ASTPropertyPathOptimizerInTest extends ASTPropertyPathOptimizer { + private int counter = 0; + + @Override + protected VarNode anonVar(final String anon) { + VarNode v = new VarNode(anon+counter++); + v.setAnonymous(true); + return v; + } + } + public AbstractOptimizerTestCase(String name) { super(name); } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestALPPinTrac773.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestALPPinTrac773.java 2014-05-04 21:33:58 UTC (rev 8191) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestALPPinTrac773.java 2014-05-04 23:19:37 UTC (rev 8192) @@ -32,7 +32,6 @@ import com.bigdata.rdf.sparql.ast.ArbitraryLengthPathNode; import com.bigdata.rdf.sparql.ast.GroupMemberNodeBase; import com.bigdata.rdf.sparql.ast.StatementPatternNode; -import com.bigdata.rdf.sparql.ast.VarNode; import com.bigdata.rdf.store.AbstractTripleStore; /** @@ -140,16 +139,7 @@ @Override IASTOptimizer newOptimizer() { return new ASTOptimizerList( - new ASTPropertyPathOptimizer() { - private int counter = 0; - - @Override - protected VarNode anonVar(final String anon) { - VarNode v = new VarNode(anon+counter++); - v.setAnonymous(true); - return v; - } - }, + new ASTPropertyPathOptimizerInTest(), new ASTRangeCountOptimizer(){ @Override Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestAll.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestAll.java 2014-05-04 21:33:58 UTC (rev 8191) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/optimizers/TestAll.java 2014-05-04 23:19:37 UTC (rev 8192) @@ -111,6 +111,8 @@ // Unit tests for enforcing bottom-up evaluation semantics. suite.addTestSuite(TestASTBottomUpOptimizer.class); + + suite.addTestSuite(TestASTPropertyPathOptimizer.class); // Unit tests for the BIGDATA "SEARCH" service optimizer. suite.addTestSuite(TestASTSearchOptimizer.class); This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <jer...@us...> - 2014-05-04 21:34:03
|
Revision: 8191 http://sourceforge.net/p/bigdata/code/8191 Author: jeremy_carroll Date: 2014-05-04 21:33:58 +0000 (Sun, 04 May 2014) Log Message: ----------- Improve the print out of debug representations - in particular ensure that the content of FILTER EXISTS and FILTER NOT EXISTS are shown prior to the optimizer step that pulls the content out as a subquery. Also shorten the names of some annotations in some debug representations. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/BOp.java branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/CoreBaseBOp.java branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/IValueExpression.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FilterNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FunctionNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IQueryNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IValueExpressionNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/StatementPatternNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/SubqueryFunctionNodeBase.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/TermNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/ValueExpressionNode.java branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/TestAST.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/BOp.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/BOp.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/BOp.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -211,6 +211,24 @@ */ boolean isController(); + /** + * The contract of this method at this level is under-specified. + * Sub-classes may choose between: + * + * - return a string representation of the object, similar to the use of {@link #toString()} + * + * Or: + * + * - return a pretty-print representation of the object with indent + * + * Note that the former contract may or may not include recursive descent through a tree-like + * object, whereas the latter almost certainly does. + * + * @param indent + * @return + */ + String toString(final int indent); + /** * Interface declaring well known annotations. * <p> Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/CoreBaseBOp.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/CoreBaseBOp.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/CoreBaseBOp.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -156,9 +156,44 @@ return sb.toString(); } + + /** + * Append a name to a string buffer, possibly shortening the name. + * The current algorithm for name shortening is to take the end of the name + * after the pen-ultimate '.'. + * @param sb + * @param longishName + */ + protected void shortenName(final StringBuilder sb, final String longishName) { + int lastDot = longishName.lastIndexOf('.'); + if (lastDot != -1) { + int lastButOneDot = longishName.lastIndexOf('.', lastDot - 1); + sb.append(longishName.substring(lastButOneDot + 1)); + return; + } + sb.append(longishName); + } + /** + * Add a string representation of annotations into a string builder. + * By default this is a non-recursive operation, however + * subclasses may override {@link #annotationValueToString(StringBuilder, BOp, int)} + * in order to make this recursive. + * @param sb + */ protected void annotationsToString(final StringBuilder sb) { - final Map<String,Object> annotations = annotations(); + annotationsToString(sb, 0); + } + + /** + * Add a string representation of annotations into a string builder. + * By default this is a non-recursive operation, however + * subclasses may override {@link #annotationValueToString(StringBuilder, BOp, int)} + * in order to make this recursive. + * @param sb + */ + protected void annotationsToString(final StringBuilder sb, final int indent) { + final Map<String,Object> annotations = annotations(); if (!annotations.isEmpty()) { sb.append("["); boolean first = true; @@ -169,20 +204,35 @@ sb.append(", "); final String key = e.getKey(); final Object val = e.getValue(); + shortenName(sb, key); + sb.append("="); if (val != null && val.getClass().isArray()) { - sb.append(key + "=" + Arrays.toString((Object[]) val)); + sb.append(Arrays.toString((Object[]) val)); } else if (key.equals(IPredicate.Annotations.FLAGS)) { - sb.append(key + "=" + Tuple.flagString((Integer) val)); + sb.append(Tuple.flagString((Integer) val)); } else if( val instanceof BOp) { - sb.append(key + "=" + ((BOp) val).toShortString()); + annotationValueToString(sb, (BOp)val, indent); } else { - sb.append(key + "=" + val); + sb.append(val); } first = false; } sb.append("]"); } - } + } + + /** + * Add a string representation of a BOp annotation value into a string builder. + * By default this is a non-recursive operation, however + * subclasses may override and give a recursive definition, which should respect + * the given indent. + * @param sb The destination buffer + * @param val The BOp to serialize + * @param indent An indent to use if a recursive approach is chosen. + */ + protected void annotationValueToString(final StringBuilder sb, final BOp val, final int indent) { + sb.append(val.toString()); + } @Override final public Object getRequiredProperty(final String name) { @@ -441,6 +491,26 @@ } + /** + * The contract of this method at this level is under-specified. + * Sub-classes may choose between: + * + * - return a string representation of the object, similar to the use of {@link #toString()} + * + * Or: + * + * - return a pretty-print representation of the object with indent + * + * Note that the former contract may or may not include recursive descent through a tree-like + * object, whereas the latter almost certainly does. + * + * @param indent + * @return + */ + public String toString(int indent) { + return toString(); + } + private static final transient String ws = " "; } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/IValueExpression.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/IValueExpression.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata/src/java/com/bigdata/bop/IValueExpression.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -26,5 +26,11 @@ * <code>null</code>. */ E get(IBindingSet bindingSet); + /** + * A string representation of a recursive structure with pretty-print indent. + * @param indent + * @return + */ + String toString(int indent); } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FilterNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FilterNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FilterNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -87,7 +87,7 @@ sb.append("\n"); sb.append(indent(indent)); - sb.append("FILTER( ").append(getValueExpressionNode()).append(" )"); + sb.append("FILTER( ").append(getValueExpressionNode().toString(indent+1)).append(" )"); // if (getQueryHints() != null) { // sb.append("\n"); Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FunctionNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FunctionNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/FunctionNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -2,6 +2,7 @@ import java.io.Serializable; import java.util.Collections; +import java.util.Iterator; import java.util.Map; import org.openrdf.model.URI; @@ -9,6 +10,7 @@ import com.bigdata.bop.BOp; import com.bigdata.bop.IValueExpression; import com.bigdata.bop.NV; +import com.bigdata.bop.BOp.Annotations; /** * AST node for anything which is neither a constant nor a variable, including @@ -295,4 +297,35 @@ } + + /** + * Provides a pretty print representation with recursive descent. + */ + @Override + public String toString(int i) { + + final StringBuilder sb = new StringBuilder(); + sb.append(getClass().getSimpleName()); + final Integer bopId = (Integer) getProperty(Annotations.BOP_ID); + if (bopId != null) { + sb.append("[" + bopId + "]"); + } + sb.append("("); + int nwritten = 0; + final Iterator<BOp> itr = argIterator(); + while(itr.hasNext()) { + final BOp t = itr.next(); + if (nwritten > 0) + sb.append(','); + if (t == null) { + sb.append("<null>"); + } else { + sb.append(((IValueExpressionNode)t).toString(i+1)); + } + nwritten++; + } + sb.append(")"); + annotationsToString(sb, i); + return sb.toString(); + } } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IQueryNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IQueryNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IQueryNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -37,8 +37,11 @@ } + /** - * Pretty print with an indent. + * A string representation of a recursive structure with pretty-print indent. + * @param indent + * @return */ String toString(final int indent); Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IValueExpressionNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IValueExpressionNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/IValueExpressionNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -34,4 +34,11 @@ */ void invalidate(); + /** + * A string representation of a recursive structure with pretty-print indent. + * @param indent + * @return + */ + String toString(final int indent); + } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/StatementPatternNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/StatementPatternNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/StatementPatternNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -574,7 +574,7 @@ if (getQueryHints() != null && !getQueryHints().isEmpty()) { sb.append("\n"); sb.append(indent(indent + 1)); - sb.append(Annotations.QUERY_HINTS); + shortenName(sb, Annotations.QUERY_HINTS); sb.append("="); sb.append(getQueryHints().toString()); } @@ -586,7 +586,7 @@ if (rangeCount != null) { sb.append("\n"); sb.append(indent(indent + 1)); - sb.append(AST2BOpBase.Annotations.ESTIMATED_CARDINALITY); + shortenName(sb, AST2BOpBase.Annotations.ESTIMATED_CARDINALITY); sb.append("="); sb.append(rangeCount.toString()); } @@ -594,7 +594,7 @@ if (keyOrder != null) { sb.append("\n"); sb.append(indent(indent + 1)); - sb.append(AST2BOpBase.Annotations.ORIGINAL_INDEX); + shortenName(sb, AST2BOpBase.Annotations.ORIGINAL_INDEX); sb.append("="); sb.append(keyOrder.toString()); } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/SubqueryFunctionNodeBase.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/SubqueryFunctionNodeBase.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/SubqueryFunctionNodeBase.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -32,6 +32,7 @@ import org.openrdf.model.URI; import com.bigdata.bop.BOp; +import com.bigdata.bop.IValueExpression; /** * A special function node for modeling value expression nodes which are @@ -119,4 +120,9 @@ } + @Override + protected void annotationValueToString(final StringBuilder sb, final BOp val, int i) { + sb.append(val.toString(i)); + } + } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/TermNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/TermNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/TermNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -88,5 +88,10 @@ return null; } + + @Override + public String toString(int i) { + return toShortString(); + } } Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/ValueExpressionNode.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/ValueExpressionNode.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/java/com/bigdata/rdf/sparql/ast/ValueExpressionNode.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -12,7 +12,7 @@ * @author <a href="mailto:tho...@us...">Bryan Thompson</a> * @version $Id$ */ -public class ValueExpressionNode extends ASTBase implements +public abstract class ValueExpressionNode extends ASTBase implements IValueExpressionNode { /** Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/TestAST.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/TestAST.java 2014-05-04 21:15:18 UTC (rev 8190) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-rdf/src/test/com/bigdata/rdf/sparql/ast/TestAST.java 2014-05-04 21:33:58 UTC (rev 8191) @@ -39,6 +39,7 @@ import com.bigdata.bop.Var; import com.bigdata.bop.ap.Predicate; import com.bigdata.journal.ITx; +import com.bigdata.rdf.internal.IV; import com.bigdata.rdf.internal.VTE; import com.bigdata.rdf.internal.constraints.CompareBOp; import com.bigdata.rdf.internal.constraints.IVValueExpression; @@ -262,7 +263,7 @@ groupBy.addExpr(new AssignmentNode(new VarNode("s"), new VarNode("s"))); final HavingNode havingBy = new HavingNode(); - havingBy.addExpr(new ValueExpressionNode(new CompareBOp(Var.var("x"), + havingBy.addExpr(new LegacyTestValueExpressionNode(new CompareBOp(Var.var("x"), Var.var("y"), CompareOp.GT))); final OrderByNode orderBy = new OrderByNode(); @@ -331,7 +332,7 @@ } public FilterNode filter(final int id) { - return new FilterNode(new ValueExpressionNode(new Filter(id))); + return new FilterNode(new LegacyTestValueExpressionNode(new Filter(id))); } public Predicate pred(final int id) { @@ -345,8 +346,23 @@ return new Filter(id); } - - private static final class Filter extends XSDBooleanIVValueExpression { + /** + * @deprecated This was just for compatibility with SOp2ASTUtility. It is + * only used by the test suite now. + */ + @Deprecated + private static final class LegacyTestValueExpressionNode extends ValueExpressionNode { + private LegacyTestValueExpressionNode(IValueExpression<? extends IV> ve) { + super(ve); + } + + @Override + public String toString(int i) { + return toShortString(); + } + } + + private static final class Filter extends XSDBooleanIVValueExpression { /** * This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 21:15:21
|
Revision: 8190 http://sourceforge.net/p/bigdata/code/8190 Author: thompsonbry Date: 2014-05-04 21:15:18 +0000 (Sun, 04 May 2014) Log Message: ----------- javadoc Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java 2014-05-04 21:14:25 UTC (rev 8189) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/AbstractLBSPolicy.java 2014-05-04 21:15:18 UTC (rev 8190) @@ -468,6 +468,9 @@ * Note: Synchronization here is used to ensure only one thread runs * this logic if the table does not exist and we get a barrage of * requests. + * + * TODO This does not sufficiently prevent against duplicate work + * for non-conditional service table update paths. */ synchronized (serviceTableRef) { This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 21:14:28
|
Revision: 8189 http://sourceforge.net/p/bigdata/code/8189 Author: thompsonbry Date: 2014-05-04 21:14:25 +0000 (Sun, 04 May 2014) Log Message: ----------- Should have been synchronized on the hostTableRef, not the local variable named 'hostTable'. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-04 21:11:13 UTC (rev 8188) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-04 21:14:25 UTC (rev 8189) @@ -439,7 +439,7 @@ * request on a service start. */ - synchronized (hostTable) { + synchronized (hostTableRef) { // Ensure that the host table exists. updateHostTable(); This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 21:11:16
|
Revision: 8188 http://sourceforge.net/p/bigdata/code/8188 Author: thompsonbry Date: 2014-05-04 21:11:13 +0000 (Sun, 04 May 2014) Log Message: ----------- Reducing WARN => INFO. Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-04 21:09:15 UTC (rev 8187) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-sails/src/java/com/bigdata/rdf/sail/webapp/lbs/policy/ganglia/GangliaLBSPolicy.java 2014-05-04 21:11:13 UTC (rev 8188) @@ -676,7 +676,8 @@ } - log.warn("hostReport=" + hostReportList); + if (log.isInfoEnabled()) + log.info("hostReport=" + hostReportList); // A dense list of host reports for our services. return hostReportList.toArray(new IHostReport[nhostsWithService]); This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 21:09:18
|
Revision: 8187 http://sourceforge.net/p/bigdata/code/8187 Author: thompsonbry Date: 2014-05-04 21:09:15 +0000 (Sun, 04 May 2014) Log Message: ----------- Rolling back the GangliaLBSPolicy default. I need to push out a release of the bigdata-ganglia.jar that declares some new methods, such as: com.bigdata.ganglia.GangliaService.getDefaultHostReportOn()[Ljava/lang/String See #624 (HA LBS) Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/WEB-INF/web.xml Modified: branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/WEB-INF/web.xml =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/WEB-INF/web.xml 2014-05-04 21:02:42 UTC (rev 8186) +++ branches/BIGDATA_RELEASE_1_3_0/bigdata-war/src/WEB-INF/web.xml 2014-05-04 21:09:15 UTC (rev 8187) @@ -139,7 +139,7 @@ <async-supported>true</async-supported> <init-param> <param-name>policy</param-name> - <param-value>com.bigdata.rdf.sail.webapp.lbs.policy.ganglia.GangliaLBSPolicy</param-value> + <param-value>com.bigdata.rdf.sail.webapp.lbs.policy.RoundRobinLBSPolicy</param-value> <description> The load balancer policy. This must be an instance of the IHALoadBalancerPolicy interface. A default policy (NOPLBSPolicy) is This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 21:02:46
|
Revision: 8186 http://sourceforge.net/p/bigdata/code/8186 Author: thompsonbry Date: 2014-05-04 21:02:42 +0000 (Sun, 04 May 2014) Log Message: ----------- Enabling platform statistics collection and ganglia listener by default for the HA deployment mode to support the GangliaLBSPolicy Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-04 20:54:42 UTC (rev 8185) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/HAJournal/HAJournal.config 2014-05-04 21:02:42 UTC (rev 8186) @@ -346,18 +346,25 @@ new NV(Journal.Options.COLLECT_QUEUE_STATISTICS, ConfigMath.getProperty("COLLECT_QUEUE_STATISTICS","false")), - // platform and process performance counters (requires external s/w on some platforms) + // Platform and process performance counters. This requires external + // software on some platforms (vmstat, pidstat, iostat, etc.). + // + // This is necessary for the GangliaLBSPolicy. new NV(Journal.Options.COLLECT_PLATFORM_STATISTICS, - ConfigMath.getProperty("COLLECT_PLATFORM_STATISTICS","false")), + ConfigMath.getProperty("COLLECT_PLATFORM_STATISTICS","true")), - // uses bigdata-ganglia module to report service metrics to ganglia. + // Use bigdata-ganglia module to build internal model of cluster load. + // + // This is required for the GangliaLBSPolicy. + new NV(com.bigdata.journal.GangliaPlugIn.Options.GANGLIA_LISTEN, + ConfigMath.getProperty("GANGLIA_LISTENER","true")), + + // Use bigdata-ganglia module to report service metrics to ganglia. + // + // This MAY be used INSTEAD of installing gmond. new NV(com.bigdata.journal.GangliaPlugIn.Options.GANGLIA_REPORT, ConfigMath.getProperty("GANGLIA_REPORT","false")), - // uses bigdata-ganglia module to build internal model of cluster load. - new NV(com.bigdata.journal.GangliaPlugIn.Options.GANGLIA_LISTEN, - ConfigMath.getProperty("GANGLIA_LISTENER","false")), - }, bigdata.kb); } This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |
From: <tho...@us...> - 2014-05-04 20:54:46
|
Revision: 8185 http://sourceforge.net/p/bigdata/code/8185 Author: thompsonbry Date: 2014-05-04 20:54:42 +0000 (Sun, 04 May 2014) Log Message: ----------- removing the jetty thread min/max overrides. They are causing problems with Property vs SystemProperty.... Modified Paths: -------------- branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices Modified: branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices =================================================================== --- branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices 2014-05-04 20:35:34 UTC (rev 8184) +++ branches/BIGDATA_RELEASE_1_3_0/src/resources/bin/startHAServices 2014-05-04 20:54:42 UTC (rev 8185) @@ -82,8 +82,6 @@ -DZK_SERVERS=${ZK_SERVERS}\ -DHA_PORT=${HA_PORT}\ "-Dcom.bigdata.hostname=${BIGDATA_HOSTNAME}"\ - "-Djetty.threads.min=${JETTY_THREADS_MIN}"\ - "-Djetty.threads.max=${JETTY_THREADS_MAX}"\ "-Djetty.port=${JETTY_PORT}"\ "-Djetty.resourceBase=${JETTY_RESOURCE_BASE}"\ "-DJETTY_XML=${JETTY_XML}"\ This was sent by the SourceForge.net collaborative development platform, the world's largest Open Source development site. |